4 RE S ULTS
We implemented the methods proposed in sec- tion 3.1, 3.2 and 3.3 and made it available for a free use on our Regen Server. It can be accessed using the link: http://regen.informatics.iupui. edu:8080/WebForBVI/index.jsp. The user in- terface looks as shown in Figure 9.
Web Filtering: We developed a generalized
Figure 9. User Interface
tool where you can enter any kind of web site’s
URL and click on filter button as shown in
Figure 10, then you will get a link for Original Web Page and Filtered Web Page as shown in Figure 11. Comparison of output generated by the system is as shown is Figure 12 (Original Page) and Figure 13 (Filtered Page).
Web Summarization: This tool also works for
Figure 10. Web Filter UI
Figure 11. Web Filter UI after clicking Filter
Button
Figure 12. Original Page
generalized web sites where you have to enter URL of the site, which first extracts the content then generates summary as well as keywords as shown in figure 14 and 15.
BVI Reader: The results of this extension can be viewed by downloading a free chrome extension on your browser and enabling it to
Figure 13. Filtered Page
Figure 14. Web Summarization UI
Figure 15. Summarized Page
play on any generalized web sites as shown in figure 16.
BVI CSR: More information about imple- mentation details can be found at our website whose user interface looks as shown in figure
For this assignment, I was allowed to improvise on a provided base code to develop a functioning web crawler. The web crawler needed to accept a starting URL and then develop a URL frontier queue of “out links” to be further explored. The crawler needed to track the number of URLs and stop adding them once the queue had reached 500 links. The crawler needed to also extract text and remove HTML tags and formatting. The assignment instructions offered using the BeautifulSoup module to achieve those goals, which I chose to do. Finally, the web crawler program needed to report metrics including the number of documents (web pages), the number of tokens extracted and processed, and the number of unique terms added to the term dictionary.
URL Filtering – URL filtering secures application servers by controlling client access. Sites that are known to be malicious or contain inappropriate content are blocked to protect the organization. TMG groups URLs into dozens of categories, including those based on security, productivity and
The first versions of WWW ((what most people call “The Web”))) provide means for people around the world to exchange information between, to work together, to communicate, and to share documentation more efficiently. Tim Berners-Lee wrote the first browser (called WWW browser) and Web server in March 1991, allowing hypertext documents to be stored, fetched, and viewed. The Web can be seen as a tremendous document store where these documents (web pages) can be fetched by typing their address into a web browser. To do that, two im- portant techniques have been developed. First, a language called Hypertext Markup Languag (HTML) tells the computers how to display documents which contain texts, photos, sounds, visuals (video), and animation, interactive
The web separating normal for the project screens Web usage on and off the framework, deters or diminishes the utilization of uses, gets screenshots each time a vigilant word is entered or read, gets all request terms, and works notwithstanding if the framework is on or off.
And then finally, within Firefox is a classic tool called Web Developer that gives a lot of data and information that anyone
RI.3.5 Use text features and search tools (e.g., key words, sidebars, hyperlinks) to locate information relevant to a given topic efficiently.
The fourth feature is a menu on the upper right hand side of the webpage
Currency is to view and evaluate when the page was last updated. Lastly, Coverage which is to compare the information with information found on other websites. Does one site provide more information, more references, and more contacts? Also compare the information on the website with information available in print sources such as books, journals, reports, etc.
In the assignment, Hands-on Activity 2A: Looking Inside your HTTP Packets (2.1,2.2.1), I learned how to send HTTP packets and a response back from the web server. The instruction of the project helped me understand the information given within an HTTP packet I create. For instance, within the project I learned how to observe and understand the information of how many HTTP packets a browser is generating when sending a request. Similar to Figure 2-20 on pg. 53 in the book Business Data Communications & Networking, my 2nd screen showed that I was using Chrome 55 as the Web browser. Within the 1st screenshot, it shows the step of how to use Rex Swain’s HTTP Viewer. Inside the box of the Rex Swain’s HTTP Viewer, it gives you the options to type in a URL, Request Type, Version, Display Format, User-Agent, Referrer, Accept-Encoding, and Auto-Follow Location. The screenshot also
Do you think schools should have web filters? In the last few years, kids were asked what they think of their schools having web filters. They all said different things such as schools should and should not have web filters. To be honest, I think web filters should not exist.
Google’s search engine allows users to input and submit data online. In return, the user would receive relevant search results. Behind the scenes upon the submission, web crawlers scan through billions of pages and link keywords from a user’s data to the publish data on the web. Their PageRank technology ranks these pages by the number and popularity of other sites that link to the page. This provides the user with accurate and popular results. Google search engines generated high revenues between advertising on its websites and selling its technology to other sites.
Launched on 15 January 2001, Wikipedia is a free encyclopedia that uses the web platform for online users to access. Boasting with over 26 million pieces of writing in 285 languages, Wikipedia has transformed to be a giant in the field of search engines optimization technology. The open source concept that it rides have made it cheap to access and a better choice for many online users. This is especially among the users who find it cumbersome to follow prolonged registration processes to access information on the internet. Any search term queried on the Google™ home page search engine will definitely give a hit from the Wikipedia site, and if not present, a prompt will request the user to create a page for such a term. In this way,
From different data sources like GKP and SEEOmoz, it shows the search volume of the queried keyword(s) and its Adwords CPC rate (Cost-per-Click). On the go you can check up to 200 keywords in a minute through this app.
Anyone who knows me knows that I care an awful lot about politics and like to stay up to date with current events. One day when I free time in study hall I chose to take advantage of the the opportunity I had to check the news on my favorite news site INFO WARS. When I did so, a sign appeared on my computer screen that read “This site has been blocked because it contains illegitimate political content”. Frustrated, I tried to go over to my second favorite news site DRUDGE REPORT, which was thankfully not blocked. However, when I clicked on an interesting article titled “Dozens killed in Belgium by suicide bomber” the same sign appeared yet again that read “This site has been blocked. Keyword: bomb”. It was extremely annoying how these sites that I had been to before and knew were not inappropriate could not be accessed. The internet filters that the school claimed was
In my opinion what makes a good web page? I have been on the net for the last, can 't really believe it, coming to ten years now. From the conception to now, I have been a regular surfer on the web, visiting thousands of pages a month, mostly in search of some thing or another. Most searches have lead to even more searches but some times I come up tops. Having been used to the all text sites of the beginning, and seeing the first implementations of more graphic based sites, there was a three-year lapse where I rarely got to surfing and missed a whole evolution in web page. When I returned in full to surfing, I discovered the face of the internet have started to change. Buttons with