Skip to content

Commit

Permalink
update setup.py, changelog, README
Browse files Browse the repository at this point in the history
  • Loading branch information
taspinar committed Nov 3, 2019
1 parent 9de8d72 commit e00fdc3
Show file tree
Hide file tree
Showing 4 changed files with 33 additions and 7 deletions.
24 changes: 19 additions & 5 deletions README.rst
Original file line number Diff line number Diff line change
@@ -1,6 +1,15 @@
Synopsis
========

|Downloads|

.. |Downloads| image:: https://pepy.tech/badge/twitterscraper
:target: https://pepy.tech/project/twitterscraper

.. |Downloads| image:: https://pepy.tech/badge/twitterscraper/month
:target: https://pepy.tech/project/twitterscraper/month


A simple script to scrape for Tweets using the Python package requests
to retrieve the content and Beautifulsoup4 to parse the retrieved
content.
Expand Down Expand Up @@ -35,19 +44,24 @@ Per Tweet it scrapes the following information:
+ Tweet-url
+ Tweet text
+ Tweet html
+ Links inside Tweet
+ Hashtags inside Tweet
+ Image URLS inside Tweet
+ Video URL inside Tweet
+ Tweet timestamp
+ Tweet Epoch timestamp
+ Tweet No. of likes
+ Tweet No. of replies
+ Tweet No. of retweets
+ Username
+ User Full Name
+ User Full Name / Screen Name
+ User ID
+ Tweet is an retweet (only when scraping for user profiles)
+ Username retweeter (only when scraping for user profiles)
+ Userid retweeter (only when scraping for user profiles)
+ Retweet ID (only when scraping for user profiles)
+ Tweet is an reply to
+ Tweet is replied to
+ List of users Tweet is an reply to
+ Tweet ID of parent tweet


In addition it can scrape for the following user information:
+ Date user joined
+ User location (if filled in)
Expand Down
12 changes: 12 additions & 0 deletions changelog.txt
Original file line number Diff line number Diff line change
@@ -1,5 +1,17 @@
# twitterscraper changelog

# 1.4.0 ( 2019-11-03 )
## Fixed
- PR228: Fixed Typo in Readme
- PR224: Force CSV quoting for all non-numeric values
## Added
- PR213: Added Dockerfile for Docker support
- PR220: Passed timeout value of 60s from method to requests.get()
- PR231: Added a lot of tweet attributes to the output, regarding links, media and replies.
- PR233: Added support for searching for the '&' sign.
## Improved
- PR223: Pretty printing the output which is dumped

# 1.3.1 ( 2019-09-07 )
## Fixed
- Change two uses of f-strings to .format() since f-strings only work well with Python 3.6+
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@

setup(
name='twitterscraper',
version='1.3.1',
version='1.4.0',
description='Tool for scraping Tweets',
url='https://github.com/taspinar/twitterscraper',
author=['Ahmet Taspinar', 'Lasse Schuirmann'],
Expand Down
2 changes: 1 addition & 1 deletion twitterscraper/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
Twitter Scraper tool
"""

__version__ = '1.3.1'
__version__ = '1.4.0'
__author__ = 'Ahmet Taspinar'
__license__ = 'MIT'

Expand Down

0 comments on commit e00fdc3

Please sign in to comment.