<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	
	>
<channel>
	<title>Comments on: Twitter Sentiment Analysis Training Corpus (Dataset)</title>
	<atom:link href="http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/feed/" rel="self" type="application/rss+xml" />
	<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/?utm_source=rss&#038;utm_medium=rss&#038;utm_campaign=twitter-sentiment-analysis-training-corpus-dataset</link>
	<description>Because the world needs another Business Intelligence blog!</description>
	<lastBuildDate>Fri, 14 Sep 2018 19:30:18 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>https://wordpress.org/?v=4.1.41</generator>
	<item>
		<title>By: Sithara Fernando</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3514</link>
		<dc:creator><![CDATA[Sithara Fernando]]></dc:creator>
		<pubDate>Thu, 17 May 2018 04:23:03 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3514</guid>
		<description><![CDATA[Dear sir ,

Can you please provide me a dataset that containing hashtags .And i need to build a hierarchy using the hashtags .I look forward to hearing from you .

Thank you.]]></description>
		<content:encoded><![CDATA[<p>Dear sir ,</p>
<p>Can you please provide me a dataset that containing hashtags .And i need to build a hierarchy using the hashtags .I look forward to hearing from you .</p>
<p>Thank you.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Sarker Monojit Asish</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3511</link>
		<dc:creator><![CDATA[Sarker Monojit Asish]]></dc:creator>
		<pubDate>Sat, 21 Apr 2018 17:07:23 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3511</guid>
		<description><![CDATA[Hi
 I am working on twitter sentiment analysis for course project.Could you send me python source code ?

Thanks]]></description>
		<content:encoded><![CDATA[<p>Hi<br />
 I am working on twitter sentiment analysis for course project.Could you send me python source code ?</p>
<p>Thanks</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: kush shrivastava</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3509</link>
		<dc:creator><![CDATA[kush shrivastava]]></dc:creator>
		<pubDate>Fri, 09 Mar 2018 11:00:00 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3509</guid>
		<description><![CDATA[Yes I too need this dataset. I have a question that how we can annotate the dataset with emotion labels?]]></description>
		<content:encoded><![CDATA[<p>Yes I too need this dataset. I have a question that how we can annotate the dataset with emotion labels?</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: mona</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3508</link>
		<dc:creator><![CDATA[mona]]></dc:creator>
		<pubDate>Thu, 22 Feb 2018 23:57:58 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3508</guid>
		<description><![CDATA[I need necessary to arabic sentment analysis dataset It was done reprocessing  before for research , please help me In the fastest time
thanks]]></description>
		<content:encoded><![CDATA[<p>I need necessary to arabic sentment analysis dataset It was done reprocessing  before for research , please help me In the fastest time<br />
thanks</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: saba</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3505</link>
		<dc:creator><![CDATA[saba]]></dc:creator>
		<pubDate>Thu, 08 Feb 2018 07:19:50 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3505</guid>
		<description><![CDATA[can u plz provide me the labelled data for spam detection in twitter]]></description>
		<content:encoded><![CDATA[<p>can u plz provide me the labelled data for spam detection in twitter</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: saba</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3504</link>
		<dc:creator><![CDATA[saba]]></dc:creator>
		<pubDate>Thu, 08 Feb 2018 07:18:50 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3504</guid>
		<description><![CDATA[can u please provide me the labelled data of twitter, as i am doing my  m.tech dessertation in twitter spam detection and i am not able to get the labelled dat can u plz provide me the same]]></description>
		<content:encoded><![CDATA[<p>can u please provide me the labelled data of twitter, as i am doing my  m.tech dessertation in twitter spam detection and i am not able to get the labelled dat can u plz provide me the same</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: The President Tweets Like a Democratic Senator &#124; lynchklablog</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3474</link>
		<dc:creator><![CDATA[The President Tweets Like a Democratic Senator &#124; lynchklablog]]></dc:creator>
		<pubDate>Fri, 26 May 2017 19:50:02 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3474</guid>
		<description><![CDATA[[&#8230;] sklearn package (MLPClassifier). For training data, I used 200,000 of the 1.5M labeled tweets from here, evenly split between positive and negative [&#8230;]]]></description>
		<content:encoded><![CDATA[<p>[&#8230;] sklearn package (MLPClassifier). For training data, I used 200,000 of the 1.5M labeled tweets from here, evenly split between positive and negative [&#8230;]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: keerti</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3469</link>
		<dc:creator><![CDATA[keerti]]></dc:creator>
		<pubDate>Sat, 22 Apr 2017 20:13:09 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3469</guid>
		<description><![CDATA[can u share me the facebook n twitter datasets for defining and predicting the human behavior in social IOT usig big data analytics]]></description>
		<content:encoded><![CDATA[<p>can u share me the facebook n twitter datasets for defining and predicting the human behavior in social IOT usig big data analytics</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Swapnil Babaji Shinde</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3461</link>
		<dc:creator><![CDATA[Swapnil Babaji Shinde]]></dc:creator>
		<pubDate>Sun, 26 Mar 2017 07:18:20 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3461</guid>
		<description><![CDATA[Please post some twitter text datasets with multiple classes e.g. sports,technology etc. for text mining]]></description>
		<content:encoded><![CDATA[<p>Please post some twitter text datasets with multiple classes e.g. sports,technology etc. for text mining</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: sopan</title>
		<link>http://thinknook.com/twitter-sentiment-analysis-training-corpus-dataset-2012-09-22/#comment-3459</link>
		<dc:creator><![CDATA[sopan]]></dc:creator>
		<pubDate>Sat, 18 Mar 2017 06:12:23 +0000</pubDate>
		<guid isPermaLink="false">http://thinknook.com/?p=710#comment-3459</guid>
		<description><![CDATA[emotion analysis dataset link plz]]></description>
		<content:encoded><![CDATA[<p>emotion analysis dataset link plz</p>
]]></content:encoded>
	</item>
</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.w3-edge.com/products/


Served from: thinknook.com @ 2026-04-22 06:36:33 by W3 Total Cache
-->