Hard disk miner motherboard 12 disks
the tools used to analyze big data mainly include open source and commercial ecosphere
open source big data ecosystem:
1. Hadoop HDFS, Hadoop maprece, HBase, hive are graally born, and the early Hadoop ecosystem is graally formed
2. Hypertable is an alternative. It exists outside the Hadoop ecosystem, but there have been some users
3. NoSQL, Membase, mongodb
business big data ecosystem:
1. All in one database / data warehouse: IBM puredata (netezza), Oracle exadata, SAP Hana, etc
2. Data warehouse: Teradata asterdata, EMC greenbloom, hpvertica, etc
3. Data Mart: qlikview, tableau, and domestic reu-bds big data
uncle's market model is just the opposite. There is no distance between us through the Internet. The rest is to rely on businesses to collect your usual search keywords, dissatisfaction with the proct, and collect and sort out ideas, and then make a model for the proct to meet the market demand, which includes the work of market research. Greatly save the development cost and the judgment of people's needs
in the past, the company did research on its own, then developed procts, and then further improved them through the opinions of users in sales
with the advent of big data, we can easily collect information through the Internet, conct research, ask and answer questions, and then make more perfect procts, that is, the proct update cycle will be greatly reced, All the unchangeable companies will die
the era of uncle's rule is the era of everyone's entrepreneurship, who finds the demand, who meets the demand can make a profit
so it's also a troubled times, with many heroes rising. It's also the age of market segmentation.
uncle is said to be just people's behavior habits. Businessmen only use the demand data part of the data
The concept of big data: big data is proposed in recent two years, which has three important characteristics: large amount of data, complex structure, and fast data update. With the development of web technology, the data generated by web users are automatically saved, and sensors are constantly collecting data. With the development of mobile Internet, the speed of automatic data collection and storage is accelerating, and the amount of data in the world is expanding. The storage and calculation of data exceeds the ability of single computer (minicomputer and mainframe), This presents a challenge to the implementation of data mining technology (generally speaking, the implementation of data mining is based on a small or large computer, or parallel computing)
The concept of data mining: data mining is an interdisciplinary subject based on the rapid development of database theory, machine learning, artificial intelligence and modern statistics, which has been applied in many fields. It involves many algorithms, such as neural network, decision tree based on machine learning, support vector machine based on statistical learning theory, classification regression tree, and association analysis. The definition of data mining is to find meaningful patterns or knowledge from massive databig data needs to be mapped into small units for calculation, and then all the results are integrated, which is the so-called map rec algorithm framework. The difference is that the original data mining technology may not be easily embedded into the map rec framework, and some algorithms need to be adjusted
the similarity or correlation between big data and data mining is that the future of data mining is no longer for a small amount of sample and random accurate data, but for massive and mixed big data. Data analysis refers to the analysis of a large amount of data collected by appropriate statistical analysis methods, It is the process of extracting useful information and forming conclusions and studying and summarizing the data in detail. This process is also the supporting process of the quality management system. In practice, data analysis can help people make judgments
extended data:
big data refers to the data set that can not be captured, managed and processed by conventional software tools within a certain period of time. It is a massive, high growth rate and diversified information asset that needs new processing mode to have stronger decision-making power, insight and process optimization ability
in "the age of big data" written by Victor Myer Schoenberg and Kenneth kukeyer, big data refers to using all data for analysis and processing instead of using the shortcut of random analysis (sampling survey). The 5V characteristics of big data (proposed by IBM): Volume (large amount), velocity (high speed), variety (diversity), value (low value density), veracity (authenticity)
with the help of the information obtained from such analysis, the company can have an in-depth understanding of all aspects of its business and customer activities, such as service usage (for metering / billing), server activities, website hits and the geographical location of equipment, personnel and physical objects, so as to respond to new situations quickly.
(answer time: January 13, 2020, in case of business change, please refer to the actual situation.)
data visualization is the most basic requirement of data analysis tools for both data analysis experts and ordinary users. Visualization can intuitively display the data, let the data speak for themselves, and let the audience hear the results
2. Data mining algorithms
visualization is for people, and data mining is for machines. Clustering, segmentation, outlier analysis and other algorithms let us go deep into the data and mine value. These algorithms not only deal with the amount of big data, but also deal with the speed of big data
3. Predictive analytical capabilities
Data Mining enables analysts to better understand data, while predictive analysis enables analysts to make some predictive judgments based on the results of visual analysis and data mining< Semantic engines (semantic engines)
the diversity of unstructured data brings new challenges to data analysis, which requires a series of tools to analyze, extract and analyze data. Semantic engine needs to be designed to intelligently extract information from "documents"
5. Data quality and master data management
data quality and data management are some of the best practices in management. Through standardized processes and tools for data processing, we can ensure a pre-defined high-quality analysis results.
2. The driver who lost the supplementary license is within the period of not submitting the physical condition Certificate (examination)
please hold the original ID card and follow & quot; Work flow of driver's license examination & quot; Apply for driver's license examination, and then apply for license supplement according to this workflow< 1. The driver holds (1) the damaged motor vehicle driver's license (the lost one will not be submitted); ② Completed motor vehicle driver's license application form; ③ Recent 1-inch bareheaded photos (white background); ④ The driver's license, ID card and (the lost is written with the loss statement) should wait in line in front of the driver's license business window in the business hall of the vehicle management office
2. The police log in to the driver's license management system to check the driver's information. ① Check whether the image of the driver on the driver's license management system is the same as the image (photo) of the person who intends to supplement the license and the image of the original and of the ID card; ② If the driver's license information has more than 12 points e to traffic violations and accidents, those with the above records will be informed to attend the study and subject theory test at the designated place (those with more than 24 points will attend the subject theory test and subject 3 actual road driving test), and they will be dealt with after passing the test; ③ If the driver's license information has traffic violations and traffic accident records, it shall be handled first and then accepted
3. If the certification materials submitted by the driver meet the requirements and the driver's information in the driver's license management system is normal, the driver's license can be made after the supplementary license information is entered, the materials are sorted out and submitted to the archive room to check with the original file
4. The driver should pay at the toll office and collect the original ID card and payment voucher at the original acceptance post 3 days later.