mattpodolak / pmaw Goto Github PK
View Code? Open in Web Editor NEWA multithread Pushshift.io API Wrapper for reddit.com comment and submission searches.
License: MIT License
A multithread Pushshift.io API Wrapper for reddit.com comment and submission searches.
License: MIT License
I've posted this issue on reddit; transfer it here as a formal issue report.
My initial requests:
posts = api.search_comments(subreddit='de', limit=None, mem_safe=True, safe_exit=True)
The metadata I obtained from this initial response.
Here is the timestamp I have:
Setting before to 1631651549
Response cache key: 186e2bb94155846df2c5d321b768b8cb
10380641 result(s) available in Pushshift
The last file checkpoint before this scraping was interrupted was:
File Checkpoint 167:: Caching 15802 Responses
Checkpoint:: Success Rate: 75.43% - Requests: 33400 - Batches: 3340 - Items Remaining: 8009761
Checkpoint:: Success Rate: 75.48% - Requests: 33500 - Batches: 3350 - Items Remaining: 8001173
Now I want to resume the process, so I plugged in the timestamp I got from the initial requests as the "before" value.
api = pmaw.PushshiftAPI()
posts = api.search_comments(subreddit="de", limit=None, mem_safe=True, safe_exit=True, before=1631651549)
However, it seems to start a new process of scraping the same query; the item remaining is what's left from the previously interrupted requests.
Loaded Cache:: Responses: 2399576 - Pending Requests: 140 - Items Remaining: 10367680
Not all PushShift shards are active. Query results may be incomplete.
Am I able to ask what is the maximum number of comments can I get from a post
#comment_id is list of string
comments_arr = pushshift_api.search_comments(ids = comment_ids)
output Error: INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 87.50% - Requests: 16 - Batches: 4 - Items Remaining: 6720
/usr/local/lib/python3.7/dist-packages/pmaw/Request.py:230: UserWarning: 6720 items were not found in Pushshift
f'{self.limit} items were not found in Pushshift')
Question: Is it true that pushshift has a delay in fetching data? Can we fetch data for current date ?
Hello!
I created a replacement service for PushShift functionality that's now restricted. See https://pullpush.io/ for details. Overall it will aim to be compatible to Dec-2022 version of PushShift API.
In case you feel burned out by the whole 3rd party tool affair; I would like to ask for permission to fork and rebrand your project :)
PullPush Actual
Again, not sure how to ask a question....but I have been using the 'search_submissions' class and it is pulling multiple duplicates for each submission. Is there something I am doing wrong?
Use parquet
instead of pickle
for caching responses.
Need to assess if this is a reasonable improvement:
parquet
cache speed for storing and retrieval against pickle
Hello Matt, Thank you for your hardwork.
I just found out that the data format of the column parent_id changed from post id, such as t3_zyw8x9, to an int , like 41497293111.
Is it still the correct parent_id? If yes, do we have a way to convert it back to the traditional post_id format?
Do I have to be a Reddit moderator to use pmaw
?
Do I have to sign up with PushShift to use pmaw
?
Hello, I'm currently developing a very simple Flask app, running only locally and I wanted to scrape some Reddit posts, using your API. I followed the example, as it's specified in the documentation, however whenever I run my script, I get the following error:
ValueError: signal only works in main thread
I read that Flask-SocketIO package causes this, but I saw that this project uses Websocket-client, which is a different package.
Would really appreciate your input.
Hi,
I found a strange issue, similar to #23. Basically, now matter what I do I get 0% Successful Rate when trying to fetch comments by their ids. I tried to pass the ids as list and as a joint string with ids delimited by ,
.
Below is the reproducible example with sample of ids that I want to fetch:
import time
from pmaw import PushshiftAPI
api = PushshiftAPI()
if __name__ == "__main__":
# list of comments
comment_ids = ['czyja5h', 'czyjfsw', 'czyjhfs', 'czyjhqz', 'czyjufm', 'czyk71o', 'czykgjk', 'czykuef', 'czyl1me', 'czylher']
# fetch the body of comments
t = time.time()
comments = api.search_comments(ids=comment_ids)
comment_list = [comment for comment in comments]
print(f"Fetching {len(comment_list)} comments complete, it took {time.time() - t:.3f}s")
print('...')
With mem_safe=True
if you don't iterate through the generator fully, responses are loaded in and add to the total length, not counting how many examples have been iterated through.
Code to recreate:
comments = api.search_comments(subreddit="science", mem_safe=True, safe_exit=True, limit=2000)
len(comments) # 2000
i = 0
comment_list = []
for c in comments:
comment_list.append(c)
i += 1
if i > 500:
break
len(comments) # 3000 at this point
len(comment_list) # 501
comment_list_2 = []
for c in comments:
comment_list_2.append(c)
len(comment_list_2) # 1499
len(comments) # 2000
I know this is possible with praw by simply saying in search() like this:
reddit.search("flair:cats")
Although, I can't find a solution when using PMAW, since paramater "q" doesnt seem to recognize the "flair:" string.
The main reasoning between flair search is, that it return much more relevant posts. For example "flair:fire", is much better than "fire" etc.
Hi!
I have been trying pmaw to request pushshift reddit data. I am trying to get all the comment ids given a post/submission id using api.search_submission_comment_ids. An example of my code is:
list_id = ['6uey5x']
comment = api.search_submission_comment_ids(ids=list_id)
However, it only returns one comment id from this requested submission id: ['dls663o']
. The request should have returned multiple comment ids according to the pushshift api: https://api.pushshift.io/reddit/submission/comment_ids/6uey5x
I also tried configuring the available options like memsafe, which didn't help. Interestingly, if I try to repeat the post id in the input, such as:
list_id = ['6uey5x', '6uey5x', '6uey5x', '6uey5x']
comment = api.search_submission_comment_ids(ids=list_id)
It gives more comment ids from the corresponding post: ['dls663o', 'dls67q6', 'dls6fgv', 'dls6fqj']
.
Could you help look into it? I feel like this workaround might have not fully utilized the pipeline.
I really apprecaite your effort and this great tool!
Test Query:
comments = api.search_comments(
after=1606262347,
before=1618581599,
subreddit="CovidVaccinated",
fields=["id","subreddit","link_id","parent_id","is_submitter","author",
"author_fullname","body","score","created_utc","permalink"],
limit=None
)
Results:
40730 result(s) available in Pushshift
Checkpoint:: Success Rate: 71.00% - Requests: 100 - Batches: 10 - Items Remaining: 33898
Checkpoint:: Success Rate: 79.00% - Requests: 200 - Batches: 20 - Items Remaining: 25661
Checkpoint:: Success Rate: 81.67% - Requests: 300 - Batches: 30 - Items Remaining: 18163
Checkpoint:: Success Rate: 81.75% - Requests: 400 - Batches: 40 - Items Remaining: 11467
Checkpoint:: Success Rate: 82.80% - Requests: 500 - Batches: 50 - Items Remaining: 4262
Checkpoint:: Success Rate: 83.02% - Requests: 583 - Batches: 60 - Items Remaining: 1
Total:: Success Rate: 83.02% - Requests: 583 - Batches: 60 - Items Remaining: 1
1 result(s) not found in Pushshift
Discovered in #12
The fact that results can be enriched with current data is really nice, but is there a way to select what values to enrich or at least exclude certain ones? For example, I want the current amount of upvotes on a post at any given time, but I don't want the updated subscriber count as I'd like to be able to see changes in it over time. Is this already a feature, or am I just doing something wrong? Also please let me know if this should be asked somewhere else.
With psaw hooked into praw I can get a comment's parent as an object by call comment.parent(). Is there a way to do this in pmaw? If not, would you consider adding it?
Using pmaw I got 0 submission returned today. It worked before. Don't know why. I exactly followed the medium post Matt wrote. I checked the status of pushshift server and it says it is fine. Any idea what has happened?
Hi, I copied the code below from the example page, it used to work but have stopped working recently, returning an empty list.
from pmaw import PushshiftAPI
api = PushshiftAPI()
comments = api.search_comments(subreddit="science", limit=1000)
comment_list = [comment for comment in comments]
comment_list
returns
[]
search_submissions also returns nothing....
Thank you so much.
Hi
When I call :
from psaw import PushshiftAPI as psawAPI
from pmaw import PushshiftAPI as pmawAPI
api = psawAPI()
api_request_generator = list(api.search_submissions(...)
And
api = pmawAPI()
api_request_generator = list(api.search_submissions(...)
How it makes search_submissions
mutithreaded? ain't both psaw and pmsaw making a single REST call to PushShift?
Hi there!
I have a very simple PMAW script:
import sys
from pmaw import PushshiftAPI
import json
import os
api = PushshiftAPI(num_workers=os.cpu_count(), rate_limit=100)
id_f = sys.argv[1]
comment_ids = [l.strip() for l in open(id_f, 'r')]
comments_arr = api.search_comments(ids=comment_ids)
for c in comments_arr:
print(json.dumps(c))
The script takes in IDs from a file, puts them into a list, and passes them into search_comments.
This script has worked fine once (en masse for lots of IDs), but for some reason, it now never completes on even a test set of 10 IDs. I'm certain I must be doing something silly, or the API has potentially changed. Could someone point me in the right direction?
Thanks!
Hi. I'm drawing blank datasets when I try to use the fields
param to filter for specific fields (in order to reduce bandwidth).
This is what I want to do:
comments = api.search_comments(subreddit=subreddit, limit=100, before=before, after=after, sort_type="score", fields="body,score,author,created_utc")
This feature is supposed to be supported by Pushshift (Pushshift: Using the fields parameter). However, when I use the PMAW wrapper to do this, it returns me a blank object.
The PMAW wrapper appears to work fine for filtering exactly one field (e.g. fields="body"
). Is it possible to filter for more fields with this wrapper?
Hi. I'm trying to use pmaw to iterate a list of usernames and scrape all their submissions and comments.
It turns out to be very slow.
Using pmaw:
gen = api2.search_submissions(author='seeellayewhy').
list(gen)
It took 7s to scrape 300+ submission
Using pmaw, it took several minutes to finish it:
from pmaw import PushshiftAPI
api = PushshiftAPI(num_workers=40, )
submisions = api.search_submissions(author='seeellayewhy', limit=None)
submission_list = [sub for sub in submisions]
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:325 result(s) available in Pushshift
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 40 - Batches: 1 - Items Remaining: 325
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 80 - Batches: 2 - Items Remaining: 323
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 120 - Batches: 3 - Items Remaining: 318
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 161 - Batches: 5 - Items Remaining: 300
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:1 result(s) not found in Pushshift
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 201 - Batches: 6 - Items Remaining: 271
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 241 - Batches: 7 - Items Remaining: 267
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 281 - Batches: 8 - Items Remaining: 252
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 321 - Batches: 9 - Items Remaining: 238
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 361 - Batches: 10 - Items Remaining: 221
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 361 - Batches: 10 - Items Remaining: 221
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
INFO:pmaw.PushshiftAPIBase:Total:: Success Rate: 100.00% - Requests: 400 - Batches: 11 - Items Remaining: 0
Is there something I'm not doing it right, or it's not recommended to do this kind of task with pmaw? From the logging, it seems like the each requests are requesting only several records? Does it has to do with search windows? If so, how to configure for this kind of tasks.
Hi Matt, I'm trying to scrape subreddit posts within a time period of six months, with a limit set to none. After irregular periods of time however, the connection gets broken apparently. Following is the code snippet and the error. I tried restarting it multiple times, but the same issue comes up. Is there any way to ensure that all data is scraped in one go? Thanks
Code :
from pmaw import PushshiftAPI
api = PushshiftAPI()
import datetime as dt
before = int(dt.datetime(2020,9,1,0,0).timestamp())
after = int(dt.datetime(2020,3,1,0,0).timestamp())
subreddit="teenagers"
posts = api.search_submissions(subreddit=subreddit, limit=None, before=before, after=after,mem_safe=True)
print(f'Retrieved {len(posts)} posts from Pushshift')
Error :
ValueError Traceback (most recent call last)
/opt/conda/lib/python3.7/site-packages/urllib3/response.py in _update_chunk_length(self)
696 try:
--> 697 self.chunk_left = int(line, 16)
698 except ValueError:ValueError: invalid literal for int() with base 16: b''
During handling of the above exception, another exception occurred:
InvalidChunkLength Traceback (most recent call last)
/opt/conda/lib/python3.7/site-packages/urllib3/response.py in _error_catcher(self)
437 try:
--> 438 yield
439/opt/conda/lib/python3.7/site-packages/urllib3/response.py in read_chunked(self, amt, decode_content)
763 while True:
--> 764 self._update_chunk_length()
765 if self.chunk_left == 0:/opt/conda/lib/python3.7/site-packages/urllib3/response.py in _update_chunk_length(self)
700 self.close()
--> 701 raise InvalidChunkLength(self, line)
702InvalidChunkLength: InvalidChunkLength(got length b'', 0 bytes read)
During handling of the above exception, another exception occurred:
ProtocolError Traceback (most recent call last)
/opt/conda/lib/python3.7/site-packages/requests/models.py in generate()
757 try:
--> 758 for chunk in self.raw.stream(chunk_size, decode_content=True):
759 yield chunk/opt/conda/lib/python3.7/site-packages/urllib3/response.py in stream(self, amt, decode_content)
571 if self.chunked and self.supports_chunked_reads():
--> 572 for line in self.read_chunked(amt, decode_content=decode_content):
573 yield line/opt/conda/lib/python3.7/site-packages/urllib3/response.py in read_chunked(self, amt, decode_content)
792 if self._original_response:
--> 793 self._original_response.close()
794/opt/conda/lib/python3.7/contextlib.py in exit(self, type, value, traceback)
129 try:
--> 130 self.gen.throw(type, value, traceback)
131 except StopIteration as exc:/opt/conda/lib/python3.7/site-packages/urllib3/response.py in _error_catcher(self)
454 # This includes IncompleteRead.
--> 455 raise ProtocolError("Connection broken: %r" % e, e)
456ProtocolError: ("Connection broken: InvalidChunkLength(got length b'', 0 bytes read)", InvalidChunkLength(got length b'', 0 bytes read))
During handling of the above exception, another exception occurred:
ChunkedEncodingError Traceback (most recent call last)
/tmp/ipykernel_19/3224342533.py in
1 subreddit="teenagers"
----> 2 posts = api.search_submissions(subreddit=subreddit, limit=None, before=before, after=after,mem_safe=True,safe_exit=True)
3 print(f'Retrieved {len(posts)} posts from Pushshift')/opt/conda/lib/python3.7/site-packages/pmaw/PushshiftAPI.py in search_submissions(self, **kwargs)
72 Response generator object
73 """
---> 74 return self._search(kind='submission', **kwargs)/opt/conda/lib/python3.7/site-packages/pmaw/PushshiftAPIBase.py in _search(self, kind, max_ids_per_request, max_results_per_request, mem_safe, search_window, dataset, safe_exit, cache_dir, filter_fn, **kwargs)
266
267 if self.req.limit > 0 and len(self.req.req_list) > 0:
--> 268 self._multithread()
269
270 self.req.save_cache()/opt/conda/lib/python3.7/site-packages/pmaw/PushshiftAPIBase.py in _multithread(self, check_total)
100 self._get, url_pay[0], url_pay[1]): url_pay for url_pay in reqs}
101
--> 102 self._futures_handler(futures, check_total)
103
104 # reset attempts if no failures/opt/conda/lib/python3.7/site-packages/pmaw/PushshiftAPIBase.py in _futures_handler(self, futures, check_total)
131 self.num_req += int(not check_total)
132 try:
--> 133 data = future.result()
134 self.num_suc += int(not check_total)
135 url = url_pay[0]/opt/conda/lib/python3.7/concurrent/futures/_base.py in result(self, timeout)
426 raise CancelledError()
427 elif self._state == FINISHED:
--> 428 return self.__get_result()
429
430 self._condition.wait(timeout)/opt/conda/lib/python3.7/concurrent/futures/_base.py in __get_result(self)
382 def __get_result(self):
383 if self._exception:
--> 384 raise self._exception
385 else:
386 return self._result/opt/conda/lib/python3.7/concurrent/futures/thread.py in run(self)
55
56 try:
---> 57 result = self.fn(*self.args, **self.kwargs)
58 except BaseException as exc:
59 self.future.set_exception(exc)/opt/conda/lib/python3.7/site-packages/pmaw/PushshiftAPIBase.py in _get(self, url, payload)
50 def _get(self, url, payload={}):
51 self._impose_rate_limit()
---> 52 r = requests.get(url, params=payload)
53 status = r.status_code
54 reason = r.reason/opt/conda/lib/python3.7/site-packages/requests/api.py in get(url, params, **kwargs)
73 """
74
---> 75 return request('get', url, params=params, **kwargs)
76
77/opt/conda/lib/python3.7/site-packages/requests/api.py in request(method, url, **kwargs)
59 # cases, and look like a memory leak in others.
60 with sessions.Session() as session:
---> 61 return session.request(method=method, url=url, **kwargs)
62
63/opt/conda/lib/python3.7/site-packages/requests/sessions.py in request(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)
540 }
541 send_kwargs.update(settings)
--> 542 resp = self.send(prep, **send_kwargs)
543
544 return resp/opt/conda/lib/python3.7/site-packages/requests/sessions.py in send(self, request, **kwargs)
695
696 if not stream:
--> 697 r.content
698
699 return r/opt/conda/lib/python3.7/site-packages/requests/models.py in content(self)
834 self._content = None
835 else:
--> 836 self._content = b''.join(self.iter_content(CONTENT_CHUNK_SIZE)) or b''
837
838 self._content_consumed = True/opt/conda/lib/python3.7/site-packages/requests/models.py in generate()
759 yield chunk
760 except ProtocolError as e:
--> 761 raise ChunkedEncodingError(e)
762 except DecodeError as e:
763 raise ContentDecodingError(e)ChunkedEncodingError: ("Connection broken: InvalidChunkLength(got length b'', 0 bytes read)", InvalidChunkLength(got length b'', 0 bytes read))
Ive been working on a project where I am gathering large amounts of comment data and I have ran into two issues with each of these functions. The first issue I ran into when using the search_submission_comment_ids function is that when searching post comment ids, there returns zero comments for any post following November 26th 2021, as well as periodically prior to this date (although I haven't done extensive testing for prior).
Following the discovery of this issue, I attempted to remedy it by checking if the use of the prior function resulted in comment data being available, and if not, then switching to the use of the search_comments function. While this did work and I was able to find comment data following the November 26th 2021 date, every API request made using the search_comments function gave a warning of "Not all Pushshift shards are active. Query Results may be incomplete.". Upon investigation using the api.metadata_.get('Shards') command, I was getting results such as:
{'failed': 0, 'skipped': 0, 'successful': 67, 'total': 74}
If anybody has any idea for why either of these issues is occurring, or why the shard metadata shows the missing shards as neither failed nor skipped and is willing to share id greatly appreciate it.
I tried the examples on README, but I can not get any result. The code and results are as shown below. I used other post_ids and can not get comment_ids either.
from pmaw import PushshiftAPI
api = PushshiftAPI()
post_ids = ['kxi2w8','kxi2g1','kxhzrl','kxhyh6','kxhwh0',
'kxhv53','kxhm7b','kxhm3s','kxhg37','kxhak9']
comment_ids = api.search_submission_comment_ids(ids=post_ids)
comment_id_list = [c_id for c_id in comment_ids]
print(len(comment_id_list))
UserWarning: 10 items were not found in Pushshift
warnings.warn(f"{self.limit} items were not found in Pushshift")
Testing is currently performed through various Jupyter notebooks on my local machine which is quite inefficient and prone to error. This could be improved by slowly adding unit tests for various components to ensure they work as expected.
There are several try catch blocks that can be improved by specifying the exceptions that are expected to occur. This will help avoid unexpected behaviour which can be caused by the wide scope of the current implementation.
Hi Matthew, thank you so much for your great work on PMAW!
I tried to use your example with a limit = 100000. It seems 0 comments will be retrieved if the limit is greater than 1000.
import datetime as dt
before = int(dt.datetime(2021,2,1,0,0).timestamp())
after = int(dt.datetime(2020,12,1,0,0).timestamp())
subreddit="wallstreetbets"
limit=100000
comments = api.search_comments(subreddit=subreddit, limit=limit, before=before, after=after)
print(f'Retrieved {len(comments)} comments from Pushshift')
The log:
WARNING:pmaw.PushshiftAPIBase:Not all PushShift shards are active. Query results may be incomplete.
Retrieved 0 comments from Pushshift
I have tried with limit = 100, 1000, 1001. It seems 0 comments will be retrieved if the limit is greater than 1000.
Can you please let me know if I missed anything? Thanks!
In psaw and in Pushshift we can pass a proxy for each request. I want to use pmaw but we use proxies heavily and hence cannot use pmaw as there is no way to pass proxies or Is there a way to do it?
PushshiftAPI(https_proxy=proxy)
it gives the smaller size and faster save\load time, while supported by the majority of data libraries
Hi! I am getting comments from the subreddit using before and after dates, but I found out that the number of unique items per day is always 100. The number of total result varies and seems right, but there are a lot of duplicates. The unique items are always 100 which is also the limit from reddit API, so I wonder if there's any connection here. Do I need to specify something in the query additionally? I tried adding size or limit but didn't seem to solve this problem (other than returning zero result when the limit is too big as others pointed out) Below is how I am sending the query now:
from pmaw import PushshiftAPI
api = PushshiftAPI()
api_request_generator = list(api.search_comments(subreddit='The_Donald',
before=calendar.timegm(until_date.timetuple()),
after=calendar.timegm(since_date.timetuple()),
safe_exit=True,
size=500,
mem_safe=True,
until=calendar.timegm(until_date.timetuple())
)
Looks great!
From the readme, am I to understand that the search_submissions
method only supports the ids
parameter?
Hello,
Thanks for v0.1.0!
I just tried it and for me it was slower than psaw. Below is a standalone that compares the completion times for both. Result: pmaw took 5m16s and psaw took 2m42s.
Am I doing something wrong?
from psaw import PushshiftAPI as psawAPI
from pmaw import PushshiftAPI as pmawAPI
import os
import time
pmaw_api = pmawAPI(num_workers=os.cpu_count()*5)
psaw_api = psawAPI()
start = time.time()
test = pmaw_api.search_submissions(after=1612372114,
before=1612501714,
subreddit='wallstreetbets',
filter=['title', 'link_flair_text', 'selftext', 'score'])
end = time.time()
print("pmaw took " + str(end - start) + " seconds.")
start = time.time()
test_gen = psaw_api.search_submissions(after=1612372114,
before=1612501714,
subreddit='wallstreetbets',
filter=['title', 'link_flair_text', 'selftext', 'score'])
test = list(test_gen)
end = time.time()
print("psaw took " + str(end - start) + " seconds.")
Can I pass a parameter to the wrapper to responses by date created? Not clear in documentaiton
I know you can get the list of comment IDs with comment_id_dict = api.search_submission_comment_ids(ids=post_ids)
and you can get a list of actual comments from a single submission ID with comments = list(api.search_comments(subreddit, link_id=post_id))
.
Any way to get a list of actual comments from a list of submission IDs?
I know I can probably for loop through the submission IDs and call api.search_comments(subreddit, link_id=post_id)
on each of them, but I'm assuming that pmaw
's multithreading magic wouldn't be used to the bests of its abilities if I did it this way.
Thanks! :)
With 1873 results available in Pushshift i get a variable enrichment between 200 and 30 items.
Does that mean that the other items were as same as Pushshift or that they were not checked?
Code below
#Imports
import praw
from pmaw import PushshiftAPI
from datetime import datetime
from datetime import timezone
import pandas as pd
reddit = praw.Reddit(
client_id="CLIENT_ID",
client_secret="CLIENT_SECRET",
password="PASSWORD",
user_agent="USER AGENT",
username="USERNAME")
api = PushshiftAPI(num_workers=20,praw=reddit)
#Set Query Parameters
datewindow_start=datetime(2021,1,1,0,0,0,tzinfo=timezone.utc)
datewindow_end=datetime(2021,1,1,0,0,0,tzinfo=timezone.utc)
query="bitcoin"
#Run Query (running only on 01/01/21 as a test)
all_dates=pd.date_range(datewindow_start,datewindow_end,freq="d")
for date in all_dates:
start_date=int(date.timestamp()-1)
end_date=int(date.timestamp()+86400)
posts=api.search_submissions(q=query,after=start_date,before=end_date)
#Dataframing
comments_df = pd.DataFrame(posts)
df_2 = comments_df[["author","subreddit","num_comments","upvote_ratio","score","title","likes","id","url"]]
df_2.head(30)
Sometimes a request within a loop of requests ends up going through hundreds (thousands?) of batches trying to locate 1 single comment. I'd rather just skip that comment and move on to the next request. Any way to implement this in the code?
example output:
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 19 - Batches: 10 - Items Remaining: 1
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 29 - Batches: 20 - Items Remaining: 1
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 39 - Batches: 30 - Items Remaining: 1
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 49 - Batches: 40 - Items Remaining: 1
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 59 - Batches: 50 - Items Remaining: 1
INFO:pmaw.PushshiftAPIBase:Checkpoint:: Success Rate: 100.00% - Requests: 69 - Batches: 60 - Items Remaining: 1
etc
etc
etc
etc
I found a strange bug, basically when calling search comments, large max_ids_per_request
will result in 0% successful rate and forever retry. Maybe the result is too long?
To reproduce:
comments = api.search_submission_comment_ids(ids="haucpf")
comment_list1 = [comment for comment in comments]
comments = api.search_comments(ids=comment_list1, max_ids_per_request=1000)
comment_list_full = [comment for comment in comments]
The above code works if we change to max_ids_per_request=500
.
Hi, there. I'm running into this error for the scraping for submissions. Could you let me know why and how can I get pass it?
subs=['depression','anxiety','suicidewatch']
for sub in subs::
posts = api.search_submissions(subreddit=sub, mem_safe=True, after=1585713600, before=1637207011, safe_exit=True)
print(f'{len(posts)} posts retrieved from Pushshift')
post_list = [post for post in posts]
pd.DataFrame(post_list).to_pickle(f"{sub}_submissions.pkl")
IndexError: list index out of range
---------------------------------------------------------------------------
IndexError Traceback (most recent call last)
<ipython-input-10-61805bcbf485> in <module>
1 for sub in subs:
----> 2 posts = api.search_submissions(subreddit=sub, mem_safe=True, after=1585713600, before=1637207011, safe_exit=True)
3 print(f'{len(posts)} posts retrieved from Pushshift')
4 post_list = [post for post in posts]
5 pd.DataFrame(post_list).to_pickle(f"{sub}_submissions.pkl")
~\AppData\Roaming\Python\Python37\site-packages\pmaw\PushshiftAPI.py in search_submissions(self, **kwargs)
72 Response generator object
73 """
---> 74 return self._search(kind='submission', **kwargs)
~\AppData\Roaming\Python\Python37\site-packages\pmaw\PushshiftAPIBase.py in _search(self, kind, max_ids_per_request, max_results_per_request, mem_safe, search_window, dataset, safe_exit, cache_dir, filter_fn, **kwargs)
261 self.req.gen_url_payloads(
262 url, self.batch_size, search_window)
--> 263
264 # check for exit signals
265 self.req.check_sigs()
~\AppData\Roaming\Python\Python37\site-packages\pmaw\PushshiftAPIBase.py in _multithread(self, check_total)
98
99 futures = {executor.submit(
--> 100 self._get, url_pay[0], url_pay[1]): url_pay for url_pay in reqs}
101
102 self._futures_handler(futures, check_total)
~\AppData\Roaming\Python\Python37\site-packages\pmaw\PushshiftAPIBase.py in _futures_handler(self, futures, check_total)
166 num = 0
167
--> 168 if num > 0:
169 # find minimum `created_utc` to set as the `before` parameter in next timeslices
170 if len(data) > 0:
IndexError: list index out of range
Why does this return 0 result(s) available in Pushshift
from pmaw import PushshiftAPI
from datetime import datetime, timedelta
api = PushshiftAPI()
seven_days_ago = (datetime.now() - timedelta(days=7)).timestamp()
comments = api.search_comments(after=seven_days_ago)
>>> 0 result(s) available in Pushshift
while if I directly query pushshift using the same parameters, I get results.
>>> seven_days_ago
>>> 1624497069.970137
https://api.pushshift.io/reddit/comment/search?after=1624497069
In addition: apparently, an integer value is needed, but only works if a subreddit or query term is given
>>> comments = api.search_comments(subreddit="science", after=seven_days_ago)
0 result(s) available in Pushshift
>>> comments = api.search_comments(subreddit="science", after=int(seven_days_ago))
15132 result(s) available in Pushshift
>>> comments = api.search_comments(q="science", after=int(seven_days_ago))
36019 result(s) available in Pushshift
Hi Matt,
Thanks for pmaw - it's a very nice library you've created!
This is not a problem with pmaw, but with my understanding of how to use it.
I have executed a search_comments with the same before and after parameters in both psaw and pmaw. It was so much faster in psaw, I was amazed! But, success rate varied from 93% to 83%, so at the end I had 40,630 comments using pmaw compared to 40,762 using psaw.
What is the best strategy for retrieving the comments that were missed? Should I assemble a list of submission ids from a search_submissions with the same parameters, based on their having num_comments greater than in the retrieved comments (or not even in the comments), then use search_submission_comment_ids with them? Or, can I utilize safe_exit, and re-run the process to see if I can get more? Or, something else?
Perhaps I it would be best to use search_submission_comment_id from the get-go? I have found that searching by id with psaw much slower that just using a date range, and as I have a range I didn't bother with it in this case. Is it slower to use than search_comments?
Cheers
Chris
Hi there, I would like to use your wonderful wrapper PMAW to conduct the following data search for all comments with the following keywords "don't hate" AND "Thomas" (i.e., any comments that contain "don't hate" and "Thomas" somewhere).
The call to the pushshift api directly works fine:
https://api.pushshift.io/reddit/comment/search?q=%22don%27t%20hate%22+Thomas
However, I cannot figure out how to translate this q="..." search query into a format that works in PMAW (or PSAW for that matter)...
For instance, the following returns 0:
comments = api.search_comments(q="%22don%27t%20hate%22+Thomas", limit=limit, before=before, after=after)
Been googling all day without any success.
in gen_url_payloads
all_ids = self.payload['ids']
if len(all_ids) == 0 and self.limit > 0:
warnings.warn(
f'{self.limit} items were not found in Pushshift')
self.limit = len(all_ids)
TypeError: '>' not supported between instances of 'NoneType' and 'int'
So seemingly limit is None type. Or I've made a mistake somehow. But I think my code is working because this only crashes intermittently (meaning it is consistent with what I would expect if Pushshift was occassionally failing).
Hi all! I was trying to restrict to a specific subreddit "web10" and I noticed that content from other subreddits with a name that includes the text "web10" is also coming up! I assume this is not the intended behavior because it could cause a lot of unrelated results to appear, especially if the subreddit name has a common word like "science" which is used by many subreddits besides r/science.
In this example (https://api.pushshift.io/reddit/search/submission?subreddit=web10), content from r/web10, r/u_Psychological-Web10, and r/u_ronaldo-web10 appears. r/u_Psychological-Web10 is a subreddit, and r/u_ronaldo-web10 is handled differently by Reddit (Reddit displays a page indicating the user has been banned, rather than a page indicating the subreddit doesn't exist) so perhaps it was previously a subreddit.
My code looks like this:
gen = list(api.search_submissions(q="fire",
sort="created_utc",
order="desc",
subreddit=subreddit,
filter=['id'],
limit=None))
for pmaw_submission in gen:
praw_submission = reddit.submission(id=pmaw_submission['id'])
print(datetime.utcfromtimestamp(praw_submission.created_utc))
Returns the following: https://pastebin.com/Y91fYYDy
You can see the years are not in order, otherwise, it kinda looks ordered, if we talking days/time only
Any idea why is this happening / how to fix this? Thanks
Hello
When i try api.search_submissions(subreddit="science", limit=None, score>=1000
I get an error SyntaxError: positional argument follows keyword argument
Do you know how to fix it ?
Thanks
Hi, is there a way to turn off the log response?
I0317 17:36:13.663935 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2852
I0317 17:36:14.613399 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2771
I0317 17:36:16.100443 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2700
I0317 17:36:16.821416 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2626
I0317 17:36:17.709720 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2555
I0317 17:36:19.522857 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2495
I0317 17:36:20.668120 139656604510016 PushshiftAPIBase.py:141] Remaining limit 2439
Hello, I've been noticing in the last couple of days that pmaw will frequently snag on batches of either submission or comment searches. It always seems to eventually pull the data, but extracting a small set of ~500 posts can take anywhere between 10 seconds and 5 minutes. Any thoughts on why?
I modified pmaw to print out the URL endpoints to see if a particular search is causing the problem, but clicking on the pushshift links shows the data in the browser incredibly fast.
I'm attaching a recent example of a pull of 500 posts with the query "test." It took a little over two minutes to pull the data. I also highlight the link with a red arrow that caused pmaw to snag for most of this request. Any thoughts on what might be causing this?
Thanks for maintaining pmaw -- it's such a great tool!
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.