Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I am trying to scrape data from two websites. I want to build a generic web parser to scrape those website. I need a generic parser because i might need to parse additional websites in the future. I want to parse the websites using java and store specific data in MySQL. What are the steps I should take? Can anyone guide me in this?
I've done this same thing for scraping the web using java and building a parser is an absolute nightmare, there are so many more things to consider that you realize at the start.
Use jSoup. It is sooo easy, and leaves you time to enjoy the finer things in life.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
I would like to create website which quizzes people on animal facts, I could do it manually, but I am wondering how I would go about automating the 'retrieval of facts' from the internet and then inputting it into the quizzes automaticallly. I'm thinking, I could build a web crawler/scraper that collects parts of information from popular sites, and then sends that data to the site, to make the quiz.
How would you go about solving this issue, and if there is a better way of achieving the same think?
PS: I am a beginner, and I might not be using the right terminology
You should take a look at Scrapy.
Scrapy is a fast high-level web crawling and web scraping framework, used to crawl websites and extract structured data from their pages. It can be used for a wide range of purposes, from data mining to monitoring and automated testing.
Here is a nice tutorial page from the documentation: Scrapy Tutorial
Look up Selenium. It is my favorite library for python. It can log into sites, click on certain elements, and you can get data very easily.
Use pip install selenium to download
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
I'm currently developing a small application for my parents' restaurant. The restaurant has a webpage where the clients have a private zone where they can see the status of their account after they log in.
Let's assume I log in succesfully and now I can see that private HTML after login. There, i'd like to parse some data.
My question is, I think, quite simple. I know I can use JSOUP for this task but I can also manage to do it by using Strings and Substrings.
Which method is best and why? I've looking for a while and I can't decide which to use.
Thanks in advance,
Alex
Jsoup is a better method, it is a library built for html parsing.
Using strings and substrings can be very error prone and lead you to writing a lot of code for something an html parsing library can already do easily.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
I'm going to work on a website with a few of my friends, but I don't really have much experience with JavaScript, HTML, or CSS.
I looked at GWT and thought it'd be good for me since my best language is Java.
Most of the friends I'm working with, however, are familiar with web development in JavaScript and will likely use it on the project.
Is it possible for me to GWT to contribute and for them to use JavaScript on the same project?
Not easily. GWT generates Javascript blocks that are intended to be self-contained, and they don't readily interoperate with other Javascript libraries, even something as ubiquitous as jQuery. In my experience, with the exception of analytics and other aspect-type scripts, you're better off picking one approach or the other and using it for everything.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
I'm looking around for a good solution for JSON-RPC either in JAVA and Javascript (both sides, but I don't mean it has to be the same project!).
Post your hints. I'd like a mainstream project with lot of examples around.
I've tried jsonrpc4j but I can't get it working with spring...
Wikipedia have list of JSON-RPC libraries in various languages.
You can use spring-json view to transfer data from client to server using JSON.
spring-json view
on the official website you will find a list of JSON-RPC libraries/Clients for all languages including java and javascript
http://json-rpc.org/wiki/implementations
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I am a student. I need to make a application. I have built the GUI. Now I need to know how to do it?
I have heard about sql server compact, but does it work together with j2me? If not, something to solve the problem?
It will be better if you can pass me a example. If you couldn't, not worry, I just to need to know a database program for mobile (with j2me).
Could it be possible without use a database program? For example, using a text file? Or any other storage mechanism.
Appreciate any help.
In j2me you can use RMS. But you can't handle with large data. Some Third party database's also available for j2me development. You can use that API's. Also refer this existing discussion on same in this forum. And look at database connection using j2me article.