Computer-Science-A-level-Ocr
-
3-3-networks8 主题
-
3-2-databases7 主题
-
3-1-compression-encryption-and-hashing4 主题
-
2-5-object-oriented-languages7 主题
-
2-4-types-of-programming-language4 主题
-
2-3-software-development5 主题
-
2-2-applications-generation6 主题
-
2-1-systems-software8 主题
-
1-3-input-output-and-storage2 主题
-
1-2-types-of-processor3 主题
-
1-1-structure-and-function-of-the-processor1 主题
-
structuring-your-responses3 主题
-
the-exam-papers2 主题
-
8-2-algorithms-for-the-main-data-structures4 主题
-
8-1-algorithms10 主题
-
7-2-computational-methods11 主题
-
7-1-programming-techniques14 主题
-
capturing-selecting-managing-and-exchanging-data
-
entity-relationship-diagrams
-
data-normalisation
-
relational-databases
-
hashing
-
symmetric-vs-asymmetric-encryption
-
run-length-encoding-and-dictionary-coding
-
lossy-and-lossless-compression
-
polymorphism-oop
-
encapsulation-oop
-
inheritance-oop
-
attributes-oop
-
methods-oop
-
objects-oop
-
capturing-selecting-managing-and-exchanging-data
-
6-5-thinking-concurrently2 主题
-
6-4-thinking-logically2 主题
-
6-3-thinking-procedurally3 主题
-
6-2-thinking-ahead1 主题
-
6-1-thinking-abstractly3 主题
-
5-2-moral-and-ethical-issues9 主题
-
5-1-computing-related-legislation4 主题
-
4-3-boolean-algebra5 主题
-
4-2-data-structures10 主题
-
4-1-data-types9 主题
-
3-4-web-technologies16 主题
-
environmental-effects
-
automated-decision-making
-
computers-in-the-workforce
-
layout-colour-paradigms-and-character-sets
-
piracy-and-offensive-communications
-
analysing-personal-information
-
monitoring-behaviour
-
censorship-and-the-internet
-
artificial-intelligence
-
the-regulation-of-investigatory-powers-act-2000
-
the-copyright-design-and-patents-act-1988
-
the-computer-misuse-act-1990
-
the-data-protection-act-1998
-
adder-circuits
-
flip-flop-circuits
-
simplifying-boolean-algebra
-
environmental-effects
search-engine-indexing
Search Engine Indexing
How do search engines work?
-
Search engines work in several stages:
-
Crawling – think of this as gathering all of the books within a library
-
Indexing – think of this as reading the books and making a structured list of the information within the books
-
Ranking – think of this as recommending books to the reader
-
Crawling
-
Web pages are discovered by search engines through software programs called crawlers (or spiders, bots, or robots)
-
Crawlers follow links from one webpage to another, systematically visiting pages on the web
-
They start from a set of seed URLs and visit other pages linked from those URLs
-
Website crawlers follow rules and guidelines established by website owners, using mechanisms like the
robots.txtfile. These guidelines direct crawlers on which areas of a website to explore or avoid, respecting website preferences and ensuring privacy -
Once a crawler reaches a webpage, it fetches the HTML content of that page
-
The crawler examines the HTML structure and retrieves information, such as text content, headings, links, and metadata
-
To understand the structure of the webpage, the HTML that was retrieved is broken down into individual components
-
This process involves identifying elements, tags, and attributes that hold valuable information like titles and headings
Indexing
-
The data extracted from the webpage is indexed, which involves storing the collected information in a structured manner within a search engine’s database
-
Each word in the document is included in the page’s index as an entry, along with the word’s position on the page
-
The index allows for quick retrieval and ranking of relevant web pages in response to user queries
Ranking
-
When a user enters a query, the search engine searches the index for matching pages and returns the results they believe are the highest quality and most relevant to the user’s query
Benefits of search engine crawling & indexing
-
The process of search engine indexing is essential for search engines to collect, examine and arrange online content
-
It involves collecting and storing information from web pages in a searchable index
-
There are many reasons for search engine crawling and indexing to happen:
-
Improved search results
-
Efficient retrieval
-
Ranking and relevance
-
Freshness and updates
-
Improved search results
-
Indexing webpages means search engines can:
-
Provide users with relevant and up-to-date search results
-
Match user searches with content which increases the chances of accurate and valuable results
-
-
This means the user is more likely to find what they’re looking for quickly, ideally on the first page of search results, without having to go to additional pages
Efficient retrieval
-
Indexing enables efficient retrieval of information
-
Search engines don’t need to scan the entire web for every search query. They can just search their indexed data to produce search results quickly
Ranking & relevance
-
Indexing enables search engines to assess the relevance and quality of web pages
-
Search result rankings are determined by various ranking algorithms that analyse indexed data. These algorithms consider factors such as keyword relevance, backlinks, and user engagement
Freshness & updates
-
Search engine crawlers periodically revisit indexed web pages to detect updates and changes
-
This process guarantees that the search results display the latest content that is currently accessible on the Internet
-
If a webpage has been updated and not re-crawled, the page may no longer be relevant for the user’s search
Responses