12/05/2022 | by Patrick Fischer, M.Sc., Founder & Data Scientist: FDS
Search engines are programs that search the Internet for relevant information. They scan all available websites for specific keywords and add the results to a list. The results are then sorted by relevance and popularity. Search engines also use several algorithms to determine the ranking of websites. These include things like the number of links to a particular page, the number of visitors to a page, and the keywords used on the page.
12/05/2022 | by Patrick Fischer, M.Sc., Founder & Data Scientist: FDS
A crawler is a program that navigates the World Wide Web. It follows links to other pages and downloads their content to analyze it. Thus, it collects information about the structure of the web and stores it in a database. For example, it can collect all available links or all words on the page. Afterwards, this database can be used for different purposes, e.g. for search engines, for price comparisons or for scientific research.
12/05/2022 | by Patrick Fischer, M.Sc., Founder & Data Scientist: FDS
Web scraping is a process that automatically extracts data from the World Wide Web. This is done by having web crawlers navigate websites and capture and store data such as text, images, videos, and other files. This process allows web developers to use collected data from the Internet to develop new and innovative ideas.
12/05/2022 | by Patrick Fischer, M.Sc., Founder & Data Scientist: FDS
Parsing is the processing of text to extract and analyze individual parts of the text. It is often used to identify and classify individual elements in a text to extract information or to understand the structure of the text. It can also be used to convert text to another format.
12/04/2022 | by Patrick Fischer, M.Sc., Founder & Data Scientist: FDS
B2B stands for business-to-business and refers to trade between companies. It is a business model in which one company sells products or services to another company or institution.