<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title><![CDATA[RSS Feed]]></title><description><![CDATA[RSS Feed]]></description><link>http://direct.ecency.com</link><image><url>http://direct.ecency.com/logo512.png</url><title>RSS Feed</title><link>http://direct.ecency.com</link></image><generator>RSS for Node</generator><lastBuildDate>Mon, 27 Apr 2026 20:28:49 GMT</lastBuildDate><atom:link href="http://direct.ecency.com/created/scrapy/rss.xml" rel="self" type="application/rss+xml"/><item><title><![CDATA[Scrapy - extracting the data you need from websites]]></title><description><![CDATA[Scrapy extracting the data you need from websites Screenshots Hunter's comment 'scraping' websites is pretty a common thing that's out there in the world. i knew companies many years back that did this]]></description><link>http://direct.ecency.com/steemhunt/@teamhumble/scrapy-extracting-the-data-you-need-from-websites</link><guid isPermaLink="true">http://direct.ecency.com/steemhunt/@teamhumble/scrapy-extracting-the-data-you-need-from-websites</guid><category><![CDATA[steemhunt]]></category><dc:creator><![CDATA[teamhumble]]></dc:creator><pubDate>Fri, 02 Nov 2018 13:11:57 GMT</pubDate><enclosure url="https://images.ecency.com/p/3jpR3paJ37V8sXC5hyVcAPad7gu98V32csqbsVNPs55rWULTiERbnT4HXfRaQavmNMYDSqVfdH2o9E2Q4x5LeXGfyeXmyapb9DJsyUVzbboLyfHzPRGE8srXwtsfQVYQizfVY?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Web crawler: A Scrapy Crawl Spider Tutorial]]></title><description><![CDATA[Have you ever had to extract lots of data from a website? There is a very simple solution called Scrapy that fits everyone’s requirements. Scrapy is a Python module that lets you easily write your own]]></description><link>http://direct.ecency.com/tutorial/@comppaz/web-crawler-a-scrapy-crawl-spider-tutorial</link><guid isPermaLink="true">http://direct.ecency.com/tutorial/@comppaz/web-crawler-a-scrapy-crawl-spider-tutorial</guid><category><![CDATA[tutorial]]></category><dc:creator><![CDATA[comppaz]]></dc:creator><pubDate>Wed, 11 Apr 2018 09:25:48 GMT</pubDate></item><item><title><![CDATA[Tutorial: How to do web scraping in Python?]]></title><description><![CDATA[When we go for data science projects, like the Titanic Survivors and Iowa House Prices projects, we need data sets to process our predictions. In above cases, those data sets have already been collected]]></description><link>http://direct.ecency.com/python/@codeastar/tutorial-how-to-do-web-scraping-in-python</link><guid isPermaLink="true">http://direct.ecency.com/python/@codeastar/tutorial-how-to-do-web-scraping-in-python</guid><category><![CDATA[python]]></category><dc:creator><![CDATA[codeastar]]></dc:creator><pubDate>Fri, 19 Jan 2018 15:27:15 GMT</pubDate><enclosure url="https://images.ecency.com/p/2gsjgna1uruvUuS7ndh9YqVwYGPLVszbFLwwpAYXZxwCpCdk3L1jLePMqHWF2x6VY8jXU8fMZZVTP9pf1NmUAW6eCMhJN7Efm3A1v42VunQDTK7d7k?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[[源代码]Python爬取网页制作电子书代码发布]]></title><description><![CDATA[最近，在GitChat发布一场Chat（Chat地址请猛戳这里），人数当天就达标了，今天把文章完成提交，同时将文章中的代码发布到码云，我就等待大家前来捧场了，Chat地址请猛戳这里。 有人爬取数据分析黄金周旅游景点，有人爬取数据分析相亲，有人大数据分析双十一，连小学生写论文都用上了大数据。]]></description><link>http://direct.ecency.com/cn/@sunsi/5e27aj-python</link><guid isPermaLink="true">http://direct.ecency.com/cn/@sunsi/5e27aj-python</guid><category><![CDATA[cn]]></category><dc:creator><![CDATA[sunsi]]></dc:creator><pubDate>Tue, 09 Jan 2018 06:00:12 GMT</pubDate></item><item><title><![CDATA[如何用 Python 爬取网页制作电子书]]></title><description><![CDATA[有人爬取数据分析黄金周旅游景点，有人爬取数据分析相亲，有人大数据分析双十一，连小学生写论文都用上了大数据。 我们每个人每天都在往网上通过微信、微博、淘宝等上传我们的个人信息，现在就连我们的钱都是放在网上，以后到强人工智能，我们连决策都要依靠网络。网上的数据就是资源和宝藏，我们需要一把铲子来挖掘它。 最近，AI 的兴起让 Python 火了一把。实际上 Python]]></description><link>http://direct.ecency.com/python/@sunsi/python</link><guid isPermaLink="true">http://direct.ecency.com/python/@sunsi/python</guid><category><![CDATA[python]]></category><dc:creator><![CDATA[sunsi]]></dc:creator><pubDate>Fri, 29 Dec 2017 00:48:00 GMT</pubDate></item><item><title><![CDATA[របៀបទាញយកអត្ថបទលើអុីនធឺណិត​ ដោយស្វ័យប្រវត្តិជាមួយនឹង Python & Scrapy]]></title><description><![CDATA[Python គឺជាភាសាកុំព្យូទ័រមួយដែលមានលក្ខណៈពិសេសគឺសំបូរទៅដោយកម្មវិធីមានស្រាប់ជាច្រើនដែលបន្ថែមនូវមុខងារផ្សេងៗធ្វើឲ្យ ភាសាមួយនេះកាន់តែមានឥទ្ធិពល និងមានភាពងាយស្រួល។ ជាក់ស្តែង Scrapy ជាកម្មវិធីមួយរបស់ Python]]></description><link>http://direct.ecency.com/cambodia/@techfree/python-and-scrapy</link><guid isPermaLink="true">http://direct.ecency.com/cambodia/@techfree/python-and-scrapy</guid><category><![CDATA[cambodia]]></category><dc:creator><![CDATA[techfree]]></dc:creator><pubDate>Tue, 21 Nov 2017 01:42:39 GMT</pubDate><enclosure url="https://images.ecency.com/p/2FFvzA2zeqoVZ5NRzV2o8MyJEzowAL6rjbt8w3dTH591Xy2buEQycYdR9UKvDRzinCU3sy5FkQbqEeFGGH4MmNQRe3kGN8JdjvVCcF9b5D36CQgdUK4fnujj1fgKg?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Have ever you wondered how you can easy get a necessary data from website?]]></title><description><![CDATA[Recently I've used Scrapy framework( It's open source project which provide a great documentation how to do it very well and fast. I can really recommend, currently I use it to scrap products prices.]]></description><link>http://direct.ecency.com/scrapy/@rafaello/have-ever-you-wondered-how-you-can-easy-get-a-necessary-data-from-website</link><guid isPermaLink="true">http://direct.ecency.com/scrapy/@rafaello/have-ever-you-wondered-how-you-can-easy-get-a-necessary-data-from-website</guid><category><![CDATA[scrapy]]></category><dc:creator><![CDATA[rafaello]]></dc:creator><pubDate>Wed, 12 Jul 2017 21:14:00 GMT</pubDate></item><item><title><![CDATA[Way to parse steemit (scraping dynamically generated frontend)]]></title><description><![CDATA[Hello!  Our today's story will be about scraping dynamically generated frontend. Current Web technologies runs a part of the code on a client (browser) side. This technologies made]]></description><link>http://direct.ecency.com/steemit/@ertinfagor/way-to-parse-steemit-scraping-dynamically-generated-frontend</link><guid isPermaLink="true">http://direct.ecency.com/steemit/@ertinfagor/way-to-parse-steemit-scraping-dynamically-generated-frontend</guid><category><![CDATA[steemit]]></category><dc:creator><![CDATA[ertinfagor]]></dc:creator><pubDate>Wed, 14 Jun 2017 09:56:15 GMT</pubDate><enclosure url="https://images.ecency.com/p/ADdPNihJzmPZxAtykknZe9Pki6wCWR4oRznph9Vz3oR7DSrfNRTrZh5J2hCLbp8Hiw2k7XdWJigvZqsmnfkorFcoR?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Scrapy - parse steemit post commenters]]></title><description><![CDATA[Recently @Inber started a contest dedicated to 700 followers. Idea of competition is that all participants must comment a post with a keyword (I`m in). At the end among participants a winner will randomly]]></description><link>http://direct.ecency.com/steemit/@ertinfagor/scrapy-parse-steemit-post-commenters</link><guid isPermaLink="true">http://direct.ecency.com/steemit/@ertinfagor/scrapy-parse-steemit-post-commenters</guid><category><![CDATA[steemit]]></category><dc:creator><![CDATA[ertinfagor]]></dc:creator><pubDate>Wed, 07 Jun 2017 11:49:51 GMT</pubDate><enclosure url="https://images.ecency.com/p/axopD2eJJx4esjUJoxc8ALaeKVwMyViifAWFU7dgQnFpBPHNrWxhd3GF8miN?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Intoduction to XPath]]></title><description><![CDATA[Hi! In this article we will research a basics of parsing web sites and extracting information. This knowledge will be necessary for us when studying Scrapy. As you know to display a web-page browser needs]]></description><link>http://direct.ecency.com/steemit/@ertinfagor/intoduction-to-xpath</link><guid isPermaLink="true">http://direct.ecency.com/steemit/@ertinfagor/intoduction-to-xpath</guid><category><![CDATA[steemit]]></category><dc:creator><![CDATA[ertinfagor]]></dc:creator><pubDate>Tue, 06 Jun 2017 10:35:18 GMT</pubDate><enclosure url="https://images.ecency.com/p/2N61tyyncFaFnNFKLegVvzmsrMAExSDXzsHdqwaiRmL2tHf8S9eA9ufF1P2Jg6AoVke8V7CJEtr5m9NbbP3p4uAKse2Weh7ndpQRrKTwhyazSUwujf4geRervJdgx3u34aaqqC5GiDEn?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Scrapy in Docker Install]]></title><description><![CDATA[Hi! Today we will install Scrapy and run a simple spider. You can find many articles about how to install in virtualenv but we install scrapy in docker container. First of all install docker. I will write]]></description><link>http://direct.ecency.com/howto/@ertinfagor/scrapy-in-docker-install</link><guid isPermaLink="true">http://direct.ecency.com/howto/@ertinfagor/scrapy-in-docker-install</guid><category><![CDATA[howto]]></category><dc:creator><![CDATA[ertinfagor]]></dc:creator><pubDate>Mon, 05 Jun 2017 11:47:06 GMT</pubDate><enclosure url="https://images.ecency.com/p/99pyU5Ga1kwr44fChspNvAzbPLnXByjawLbsmEfvhq2Vw1Fq7ZfjiEBfsVspsQ4KtpZ5ed1JfwwnMJnSmztPJBpxdcSWZj4CBuEzUNvPwN8kXnaE41DZP6iPXrPa2XhTM1?format=match&amp;mode=fit" length="0" type="false"/></item><item><title><![CDATA[Python Scrapy]]></title><description><![CDATA[Image credit Hi!  Today I want to write an article on Data Science theme. One of the first tasks of Data Science is a ETL (Extract Transform Load) and I want to tell about Scrapy. Scrapy is a Python]]></description><link>http://direct.ecency.com/howto/@ertinfagor/python-scrapy</link><guid isPermaLink="true">http://direct.ecency.com/howto/@ertinfagor/python-scrapy</guid><category><![CDATA[howto]]></category><dc:creator><![CDATA[ertinfagor]]></dc:creator><pubDate>Sat, 03 Jun 2017 19:02:06 GMT</pubDate><enclosure url="https://images.ecency.com/p/x7L2VSNEiyAFMrpiG2ns3CB2gK32YGyd3PzYWd5t2qpCdo6bect8Mceakn4wQhEiyJBt6dt5cAGb3eW?format=match&amp;mode=fit" length="0" type="false"/></item></channel></rss>