Granthaalayah
CRIMINALITY DATA SCRUTINY USING LOGISTIC REGRESSION ALGORITHM

Criminality data scrutiny using logistic regression algorithm

 

Dr. G. Soma Sekhar 1, Puvvada Abhinaya 2, Rachapudi Jayani 2, Dr. D. Srinivasa Rao 3

 

1 Associate Professor, Department of CSE, Geethanjali College of Engineering and Technology, Hyderabad, India

2 Department of CSE, Geethanjali College of Engineering and Technology, Hyderabad, India

3 Assistant Professor, Department of CSE, GITAM Deemed to be University, Hyderabad, India

 

A picture containing logo

Description automatically generated

ABSTRACT

The quantity of criminal cases in India is rising rapidly, which is the reason there are likewise a rising number of cases as yet extraordinary. Criminal cases are expanding ceaselessly, making it difficult to sort and determine them. It's essential to perceive an area's patterns of crime to prevent it from working out. On the off chance that the specialists entrusted with researching violations have a strong comprehension of the patterns in crime happening in a specific area, they will actually want to improve. Finding the examples of crime in a particular area should be possible by applying AI and different calculations. This review predicts the sorts of wrongdoings that will happen in a given area utilizing wrongdoing information, which facilitates the characterization of criminal cases and considers suitable activity. This exploration utilizes information from the most recent 18 years that were assembled from various solid sources. This article utilized choice, erasing invalid qualities, and mark encoding to clean and support the information since information pre-handling is similarly just about as vital as definite expectation. A compelling AI model for estimating the resulting criminal case is given by this exploration.

 

Received 24 May 2023

Accepted 25 June 2023

Published 10 July 2023

Corresponding Author

Dr. G. Soma Sekhar, gsomasekharonline@gmail.com

DOI 10.29121/granthaalayah.v11.i6.2023.5182000  

Funding: This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.

Copyright: © 2023 The Author(s). This work is licensed under a Creative Commons Attribution 4.0 International License.

With the license CC-BY, authors retain the copyright, allowing anyone to download, reuse, re-print, modify, distribute, and/or copy their contribution. The work must be properly attributed to its author.

 

Keywords: Data Scrutiny, Criminal Cases, AI Model


1. INTRODUCTION

As additional crimes are being carried out, there are as of now more crook cases forthcoming in India than at any time in recent memory. A total interior examination and examination ought to be led to determine a case in light of explicit information. Given the volume of wrongdoing information now accessible in India, it is excessively extreme for the authorities to break down and settle on these lawbreaker cases. This study recognizes this as a critical issue and spotlights thinking of an answer for how violations are chosen to be committed. The logical field of AI manages PC decision-production without human inclusion. Mechanized or self-driving vehicles are contemporary instances of how AI is being utilized in many fields. We can figure explicit results in light of our bits of feedback and propose an answer for tackling wrongdoing cases in India by utilizing AI calculations. Grouping and relapse are the two most utilized classes of expectation calculations. This area of wrongdoing information forecast utilizes grouping. Order is a directed expectation technique that has been applied in different fields, like stock estimating and the drug business. This paper's significant objective is to look at a couple of calculations that might be utilized to gauge and examine wrongdoing information and to expand the accuracy of such models through information handling to come by improved results. The objective is to approve the preparation dataset with the test dataset to foster the fundamental model to anticipate the information. Here, choice trees, irregular backwoods, and strategic relapse displaying are being utilized for order.

 

2. Logistic regression algorithm

Logistic regression is a statistical approach employed in the development of machine learning models, wherein the dependent variable takes on a dichotomous form, meaning it assumes binary values. Logistic regression is a statistical method employed to elucidate the data and the linkage relationship between the response variable and at least one predictor variable. The independent variables in a research study may represent nominal, ordinal, or interval data. The term "logistic regression" derives from the utilization of the logistic function. The sigmoid function, commonly referred to as the logistic function, is utilized in mathematical modeling and analysis. The logistic function exhibits a range of values that lies within the interval of 0 and 1.

A graph of a logistic regression

Description automatically generated with low confidence

 

3. RELATED WORK

McClendon, Lawrence, and Natarajan McClendon & Meghanathan (2015) proposed an undertaking "Utilizing AI calculations to examine wrongdoing information." which manages the information mining and AI methods to recognize wrongdoing and forestall it. Their exploration was utilized to feature WEKA, a non-proprietary information mining programming, to direct a similar report amongst vicious wrongdoing designs from the Networks and Wrongdoing Unnormalized Dataset given by the College of California-Irvine storehouse and genuine wrongdoing factual information in order to achieve province of Mississippi that has been given by neighborhoodscout.com. They carried out Straight Relapse, added substance Relapse, and Choice Stump calculations utilizing a similar limited set of elements, on the Networks and Wrongdoing Dataset. In general, the direct relapse calculation played out the best among the three chosen calculations. The extent of their undertaking is to demonstrate the way that compelling and exact AI calculations can point towards information.

Bharati & Sarvanaguru (2018) proposed an examination of “Crime Prediction and Analysis Using Machine Learning” in the International Research Journal of Engineering and Technology (IRJET). Their venture is spurred by the way that India has seen an ascent in criminal cases because of which the quantity of cases forthcoming is likewise stacking up. It is essential to first comprehend the wrongdoing design to keep it from happening. This paper utilizes a wrongdoing informational index and forecasts the sorts of violations in a specific region which helps in setting up the grouping of criminal cases. The informational index utilized is gathered for a long time from confided-in sources. This paper features the utilization of element choice, eliminating invalid qualities and name encoding to clean the information. The principal point of this paper is to consider a few calculations to foresee and examine the wrongdoing information and furthermore work on the exactness of those models by information handling to get improved results. The machine learning models used here for prediction are Logistic Regression, Decision Tree Classification and Random Forest Classification.

Wang et al. (2013) presented their exploration of "Learning to detect patterns of Crime." at the Joint European Gathering on AI and information revelation in Data sets. Springer, Berlin, Heidelberg. They present a novel, vigorous information-driven regularization methodology called Adaptive Regularized Boosting (AR-Lift), persuaded by a craving to decrease over-expungement of data. They supplant AdaBoost's complex system with a standardized delicate edge that compromises between a bigger edge, to the detriment of misclassified errors. Limiting this regularized remarkable misfortune brings about a helping calculation that loosens up the frail learning supposition further: it can utilize classifiers with mistakes more prominent. This empowers a characteristic expansion to multiclass aiding and further decreases overfitting in both double and multiclass cases. They infer limits for preparing and speculation mistakes and relate them to AdaBoost. At last, they show observational outcomes and crucial information that lay out the power of our methodology and further developed execution generally.

Sathyadevan et al. (2014) presented their exploration of “Crime analysis and prediction using data mining” at the First International Conference on Networks & Soft Computing (ICNSC 2014) 2014 Aug 19 (pp. 406-412) IEEE. Wrongdoing examination and counteraction is a deliberate methodology for recognizing and analyzing fraud examples and patterns. This architecture can foresee locales that have a high probability of wrongdoing events and can imagine crisis regions. The rising appearance of mechanized frameworks, wrongdoing information experts can help Policemen to accelerate the most common way of tackling violations. The idea of information mining allows us to separate ambiguous and useful data from unstructured information. Here, software engineering and law enforcement foster an information gathering technique that can assist with settling violations quicker. Rather than zeroing in on reasons for wrongdoing events like the crook foundation of the criminal, political hostility, and are zeroing in for the most part on wrongdoing elements every day.

 

 

4. PROPOSED FRAME WORK

The review work, which was finished by glancing through many such records, filled in as the establishment for the proposed framework. The area and the kinds of wrongdoing that are going on there are utilized to gauge practically all violations. Straight Relapse, Choice Tree, and Irregular Woodland models are utilized in this paper to foresee wrongdoings since research in the past has shown that they will often have great exactness. This study uses the from data.world.com dataset. As per the state and year, the informational index incorporates different wrongdoing sorts that are executed in India. This exposition involves a few wrongdoing sorts as its feedback and results in the geographic district where every wrongdoing is perpetrated. Information cleaning, highlight determination, eliminating invalid qualities, and information scaling through normalizing and normalizing are all essential for the information pre-handling process. During information planning, the information is liberated from invalid qualities that could decisively impact the model's precision, and component determination is utilized to pick just the fundamental attributes that won't influence the model's exactness. The picked models, for example, Strategic Relapse, Choice Tree, and Irregular Backwoods, are prepared by partitioning the information into train and test many figures pre-handling. We use grouping models for this situation since the result that is required is an all-out esteem. The information forecast is finished utilizing the Python language.

 

5. Implementation and Result Analysis

 

A screen shot of a graph

Description automatically generated with medium confidence

 

A screenshot of a graph

Description automatically generated with medium confidence

 

A screenshot of a graph

Description automatically generated with low confidence

 

A map of the united states

Description automatically generated with medium confidence

 

Test case Id

Test case

Name

Test case

description

Test

Steps

Test Case Status

Test

Priority

Step

Expected

Actual

1

Start the Application

Host the application and test if it starts making sure the required software is available

If it does not start

We cannot run the application

The application hosts success

High

High

2

Home page

Check the deployment environment for properly loading the application

If it does not load

We cannot access the application

The application is running successfully

High

High

3

User Mode

Verify the working of the application in freestyle mode

If it does not respond

We cannot use the freestyle mode

The application displays the freestyle page

High

High

4

Data Input

Verify if the application takes input and updates

If it fails to take the input or store in the database

We cannot proceed further

The application updates the input to application

High

High

 

6. Conclusion

Essential data on criminal conduct in an area clearly has markers that AI specialists will use to classify a crime given an area and date. The dataset's unequal classes cause issues for the preparation specialist, yet it was inclined to fix the issue by finishing and under-inspecting the dataset. This paper confers a wrongdoing information expectation involving a Colab journal with Python as its center language. Python gives work in libraries like Pandas and NumPy along which the effort will be finished all the more rapidly, and Scikit gives every one of the cycles of how to utilize various libraries given by python. The exactness of the Arbitrary Woods not entirely set in stone to be great with a precision of 95.122%. Expectation results shift for various calculations.

 

CONFLICT OF INTERESTS

None. 

 

ACKNOWLEDGMENTS

None.

 

REFERENCES

Bharati, A., & Sarvanaguru, R. A. K. (2018). Crime Prediction Andanalysis Using Machine Learning. In International Research Journal of Engineering and Technology (IRJET), 05(09), 1037-1042.  

Chen, H., Chung, W., Xu, J. J., Wang, G., Qin, Y., and Chau, M. (2004). Crime Data Mining : A General Framework and Some Examples. Computer, 37(4), 50-56. https://doi.org/10.1109/MC.2004.1297301.

McClendon, L., & Meghanathan, N. (2015). Using Machine Learning Algorithms to Analyze Crime Data. Machine Learning and Applications, 2(1), 1-12. https://doi.org/10.5121/mlaij.2015.2101.

Nath, S. V. (2006). Crime Pattern Detection Using Data Mining. In. 2006 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology Workshops, 41-44. https://doi.org/10.1109/WI-IATW.2006.55.

Sathyadevan, S., Devan, M. S., & Gangadharan, S. S. (2014). Crime Analysis and Prediction Using Data Mining First International Conference on Networks Y Soft Computing (ICNSC2014). IEEE Publications, 406-412. https://doi.org/10.1109/CNSC.2014.6906719.

Wang, T., Rudin, C., Wagner, D., & Sevieri, R. (2013). Learning to Detect Patterns of Crime. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 515-530. https://doi.org/10.1007/978-3-642-40994-3_33.

     

 

 

 

 

Creative Commons Licence This work is licensed under a: Creative Commons Attribution 4.0 International License

© Granthaalayah 2014-2023. All Rights Reserved.