1
|
|
|
#!/usr/bin/env python |
2
|
|
|
""" |
3
|
|
|
Wordnik and Twitter utility functions |
4
|
|
|
""" |
5
|
|
|
|
6
|
|
|
# ================ GENERAL ================== |
7
|
|
|
|
8
|
|
|
import argparse |
9
|
|
|
try: |
10
|
|
|
import ConfigParser as configparser |
11
|
|
|
except ImportError: |
12
|
|
|
import configparser |
13
|
|
|
import csv |
14
|
|
|
import datetime |
15
|
|
|
import os |
16
|
|
|
import random |
17
|
|
|
import re |
18
|
|
|
import time |
19
|
|
|
|
20
|
|
|
from wordnik import swagger, AccountApi, WordListApi |
21
|
|
|
from twitter import Twitter, OAuth # pip install twitter |
22
|
|
|
|
23
|
|
|
# For Python 2.x |
24
|
|
|
try: |
25
|
|
|
input = raw_input |
26
|
|
|
except NameError: |
27
|
|
|
pass |
28
|
|
|
|
29
|
|
|
# Log time |
30
|
|
|
print(time.ctime()) |
31
|
|
|
|
32
|
|
|
|
33
|
|
|
# Test mode doesn't actually save csv, ini or update Wordnik or Twitter |
34
|
|
|
TEST_MODE = False |
35
|
|
|
|
36
|
|
|
TWEET_CHOICES = ( |
37
|
|
|
'none', # 'none' must be first |
38
|
|
|
'latest', 'latest_onetweet', '24hours', '7days', '30days', 'thisyear', |
39
|
|
|
'alltime', |
40
|
|
|
'retweet', 'random') # 'retweet' and 'random' must be last |
41
|
|
|
|
42
|
|
|
DAY_IN_SECONDS = 24 * 60 * 60 |
43
|
|
|
|
44
|
|
|
|
45
|
|
|
# Remove duplicates from a list but keep in order |
46
|
|
|
# http://stackoverflow.com/questions/480214/ |
47
|
|
|
def dedupe(seq): |
48
|
|
|
seen = set() |
49
|
|
|
seen_add = seen.add |
50
|
|
|
return [x for x in seq if x not in seen and not seen_add(x)] |
51
|
|
|
|
52
|
|
|
|
53
|
|
|
# cmd.exe cannot do Unicode so encode first |
54
|
|
|
def print_it(text): |
55
|
|
|
print(text.encode('utf-8')) |
56
|
|
|
|
57
|
|
|
|
58
|
|
|
def do_argparse(description=None): |
59
|
|
|
parser = argparse.ArgumentParser( |
60
|
|
|
description=description, |
61
|
|
|
formatter_class=argparse.ArgumentDefaultsHelpFormatter) |
62
|
|
|
parser.add_argument( |
63
|
|
|
'-t', '--tweet', default='latest', choices=TWEET_CHOICES, |
64
|
|
|
help="How to tweet the results.") |
65
|
|
|
return parser |
66
|
|
|
|
67
|
|
|
# The `stuff` list looks like: |
68
|
|
|
# [ |
69
|
|
|
# ["I love the word", "I hate the word"], # search term |
70
|
|
|
# [love_max_id, hate_max_id], |
71
|
|
|
# ["twitter-loves", "twitter-hates"] # Wordnik word list permalink |
72
|
|
|
# ] |
73
|
|
|
|
74
|
|
|
|
75
|
|
|
def load_ini(ini_file, stuff): |
76
|
|
|
config = configparser.ConfigParser() |
77
|
|
|
result = config.read(ini_file) |
78
|
|
|
if result: |
79
|
|
|
# print(summary) |
80
|
|
|
for i in range(len(stuff[1])): |
81
|
|
|
# Load max IDs using permalink as key |
82
|
|
|
stuff[1][i] = config.get("max_ids", stuff[2][i]) |
83
|
|
|
return stuff |
84
|
|
|
|
85
|
|
|
print("Loaded: " + stuff[1]) |
86
|
|
|
|
87
|
|
|
|
88
|
|
|
def save_ini(ini_file, stuff): |
89
|
|
|
print("Save: " + str(stuff[1])) |
90
|
|
|
|
91
|
|
|
config = configparser.ConfigParser() |
92
|
|
|
config.add_section("max_ids") |
93
|
|
|
for i in range(len(stuff[1])): |
94
|
|
|
# Save max IDs using permalink as key |
95
|
|
|
config.set("max_ids", stuff[2][i], stuff[1][i]) |
96
|
|
|
|
97
|
|
|
if TEST_MODE: |
98
|
|
|
return |
99
|
|
|
with open(ini_file, 'wb') as configfile: |
100
|
|
|
config.write(configfile) |
101
|
|
|
|
102
|
|
|
|
103
|
|
|
def update_csv(csv_file, search_term, words, statuses): |
104
|
|
|
file_exists = os.path.exists(csv_file) |
105
|
|
|
if TEST_MODE: |
106
|
|
|
return |
107
|
|
|
fd = open(csv_file, 'ab') |
108
|
|
|
try: |
109
|
|
|
writer = csv.writer(fd) |
110
|
|
|
if not file_exists: # add header |
111
|
|
|
writer.writerow(( |
112
|
|
|
'word', 'search_term', 'created_at', 'id_str', |
113
|
|
|
'screen_name', 'user_name', 'text')) |
114
|
|
|
for i, status in enumerate(statuses): |
115
|
|
|
csv_data = [ |
116
|
|
|
words[i], search_term, status['created_at'], status['id_str'], |
117
|
|
|
status['user']['screen_name'], status['user']['name'], |
118
|
|
|
status['text'].replace('\n', ' ')] |
119
|
|
|
for i, field in enumerate(csv_data): |
120
|
|
|
csv_data[i] = field.encode('utf8') |
121
|
|
|
writer.writerow(csv_data) |
122
|
|
|
|
123
|
|
|
finally: |
124
|
|
|
fd.close() |
125
|
|
|
|
126
|
|
|
|
127
|
|
|
def get_pattern(search_term, target_word_follows_search_term): |
128
|
|
|
# word boundary, one or more word chars, any '-*, |
129
|
|
|
# one or more word chars, word boundary |
130
|
|
|
word_pattern = "[^\w]*(\w+(['-\*]*\w)*)[^\w]*" |
131
|
|
|
|
132
|
|
|
if target_word_follows_search_term: |
133
|
|
|
# Matches search term ("I love the word") |
134
|
|
|
# followed by whitespace then at least one |
135
|
|
|
# [whitespace, period, exclamation mark, comma, |
136
|
|
|
# brackets, question mark] |
137
|
|
|
# pattern = re.compile( |
138
|
|
|
# search_term + "\s+([^\s.!,()?]+)", re.IGNORECASE) |
139
|
|
|
|
140
|
|
|
# \s = whitespace |
141
|
|
|
# \w = word characters (a-zA-Z0-9_) but re.UNICODE allows umlauts |
142
|
|
|
# and things search term, whitespace, then any number of non-word |
143
|
|
|
# chars, then begin group: one or more word chars, then any number |
144
|
|
|
# of apostrophes and hyphens as long as they are followed by a word |
145
|
|
|
# char. Then end the group with any number of non-word chars. |
146
|
|
|
pattern = re.compile( |
147
|
|
|
search_term + "\s+" + word_pattern, |
148
|
|
|
re.IGNORECASE | re.UNICODE) |
149
|
|
|
else: |
150
|
|
|
# Matches at least something that's NOT |
151
|
|
|
# [whitespace, period, exclamation mark, comma, |
152
|
|
|
# open bracket, close bracket], |
153
|
|
|
# followed by at least one |
154
|
|
|
# [whitespace, period, exclamation mark, comma] |
155
|
|
|
# and then "is my new etc." |
156
|
|
|
pattern = re.compile( |
157
|
|
|
word_pattern + "\s+" + search_term, |
158
|
|
|
re.IGNORECASE | re.UNICODE) |
159
|
|
|
|
160
|
|
|
return pattern |
161
|
|
|
|
162
|
|
|
|
163
|
|
|
def word_from_text(text, pattern, search_term): |
164
|
|
|
""" If matching word found in tweet text, return it. Else return None """ |
165
|
|
|
print_it(text) |
166
|
|
|
|
167
|
|
|
# Ignore retweets |
168
|
|
|
if text.startswith('RT'): |
169
|
|
|
return None |
170
|
|
|
|
171
|
|
|
# Ignore retweets |
172
|
|
|
if ' RT ' in text and text.find(' RT ') < text.find(search_term): |
173
|
|
|
return None |
174
|
|
|
|
175
|
|
|
# Ignore tweets beginning with a curly left double quote, |
176
|
|
|
# they're often quoting another person's tweet |
177
|
|
|
if text[0] == u"\u201c": |
178
|
|
|
return None |
179
|
|
|
|
180
|
|
|
# Ignore, probably quoting another's tweet |
181
|
|
|
# (but don't ignore: '"word" is my new favourite') |
182
|
|
|
if text.startswith('"@'): |
183
|
|
|
return None |
184
|
|
|
|
185
|
|
|
match = re.search(pattern, text) |
186
|
|
|
if match: |
187
|
|
|
word = match.group(1).lower() |
188
|
|
|
|
189
|
|
|
if len(word) == 0: |
190
|
|
|
return None |
191
|
|
|
|
192
|
|
|
# Ignore some common words |
193
|
|
|
if word.lower() in [ |
194
|
|
|
"it", "this", "that", "which", "and", |
195
|
|
|
"a", "of", "in", "but", "there"]: |
196
|
|
|
return None |
197
|
|
|
|
198
|
|
|
# Ignore if any unbalanced brackets |
199
|
|
|
open = 0 |
200
|
|
|
for char in word: |
201
|
|
|
if char == "(": |
202
|
|
|
open += 1 |
203
|
|
|
elif char == ")": |
204
|
|
|
open -= 1 |
205
|
|
|
if open < 0: |
206
|
|
|
return None |
207
|
|
|
if open != 0: |
208
|
|
|
return None |
209
|
|
|
|
210
|
|
|
# OK, got something |
211
|
|
|
print_it(">" + word + "<") |
212
|
|
|
return word |
213
|
|
|
|
214
|
|
|
# Nothing found |
215
|
|
|
return None |
216
|
|
|
|
217
|
|
|
|
218
|
|
|
def extract_words(search_term, target_word_follows_search_term, results): |
219
|
|
|
words = [] |
220
|
|
|
statuses = [] |
221
|
|
|
pattern = get_pattern(search_term, target_word_follows_search_term) |
222
|
|
|
|
223
|
|
|
for status in results['statuses']: |
224
|
|
|
# Ignore a Twitter bot |
225
|
|
|
print(status['user']['screen_name']) |
226
|
|
|
print(status['user']['screen_name'] == "unrepedant") |
227
|
|
|
print(type(status['user']['screen_name']), type("unrepedant")) |
228
|
|
|
if status['user']['screen_name'] == "unrepedant": |
229
|
|
|
continue |
230
|
|
|
|
231
|
|
|
text = status['text'] |
232
|
|
|
print("----") |
233
|
|
|
|
234
|
|
|
word = word_from_text(text, pattern, search_term) |
235
|
|
|
if word is not None: |
236
|
|
|
# print_it(status['user']['screen_name']) |
237
|
|
|
words.append(word) |
238
|
|
|
statuses.append(status) |
239
|
|
|
|
240
|
|
|
return words, statuses |
241
|
|
|
|
242
|
|
|
|
243
|
|
|
def find_words( |
244
|
|
|
search_term, target_word_follows_search_term, results, csv_file): |
245
|
|
|
words, statuses = extract_words( |
246
|
|
|
search_term, target_word_follows_search_term, results) |
247
|
|
|
update_csv(csv_file, search_term, words, statuses) |
248
|
|
|
return words |
249
|
|
|
|
250
|
|
|
|
251
|
|
|
def find_colnum(heading, row): |
252
|
|
|
"""Find the coloumn number for a given heading""" |
253
|
|
|
# Find word column |
254
|
|
|
found_colnum = None |
255
|
|
|
for colnum, col in enumerate(row): |
256
|
|
|
if heading == col: |
257
|
|
|
found_colnum = colnum |
258
|
|
|
break |
259
|
|
|
return found_colnum |
260
|
|
|
|
261
|
|
|
|
262
|
|
|
def words_and_ids_from_csv(csv_file, search_term, seconds_delta=None): |
263
|
|
|
"""Load the CSV and return a random ID from the given time period""" |
264
|
|
|
cutoff = 0 |
265
|
|
|
if seconds_delta: |
266
|
|
|
epoch_time = int(time.time()) |
267
|
|
|
cutoff = epoch_time - seconds_delta |
268
|
|
|
|
269
|
|
|
word_colnum, searchterm_colnum, created_at_colnum = None, None, None |
270
|
|
|
matched_words, eligible_ids = [], [] |
271
|
|
|
seen = set() # avoid duplicates |
272
|
|
|
ifile = open(csv_file, "r") |
273
|
|
|
reader = csv.reader(ifile) |
274
|
|
|
|
275
|
|
|
for rownum, row in enumerate(reader): |
276
|
|
|
# Save header row |
277
|
|
|
if rownum == 0: |
278
|
|
|
# Find columns |
279
|
|
|
word_colnum = find_colnum("word", row) |
280
|
|
|
searchterm_colnum = find_colnum("search_term", row) |
281
|
|
|
created_at_colnum = find_colnum("created_at", row) |
282
|
|
|
text_colnum = find_colnum("text", row) |
283
|
|
|
id_str_colnum = find_colnum("id_str", row) |
284
|
|
|
|
285
|
|
|
else: # not header |
286
|
|
|
if not row: |
287
|
|
|
continue |
288
|
|
|
# Avoid duplicates |
289
|
|
|
if row[id_str_colnum] in seen: |
290
|
|
|
continue |
291
|
|
|
seen.add(row[id_str_colnum]) |
292
|
|
|
|
293
|
|
|
# Kill the spambot! |
294
|
|
|
if row[searchterm_colnum] != search_term: |
295
|
|
|
continue |
296
|
|
|
text = row[text_colnum] |
297
|
|
|
if text[0] == "@" and \ |
298
|
|
|
"I love the word douchebag. http://t.co/" in text: |
299
|
|
|
# print(row[text_colnum]) |
300
|
|
|
continue |
301
|
|
|
|
302
|
|
|
# seconds since epoch: |
303
|
|
|
timestamp = time.mktime(time.strptime( |
304
|
|
|
row[created_at_colnum], '%a %b %d %H:%M:%S +0000 %Y')) |
305
|
|
|
if timestamp > cutoff: |
306
|
|
|
eligible_ids.append(row[id_str_colnum]) |
307
|
|
|
matched_words.append(row[word_colnum].decode('utf-8')) |
308
|
|
|
|
309
|
|
|
ifile.close() |
310
|
|
|
|
311
|
|
|
return eligible_ids, matched_words |
312
|
|
|
|
313
|
|
|
|
314
|
|
|
def pick_a_random_tweet(csv_file, search_term, seconds_delta=None): |
315
|
|
|
"""Load the CSV and return a random ID from the given time period""" |
316
|
|
|
|
317
|
|
|
eligible_ids, matched_words = words_and_ids_from_csv(csv_file, search_term, |
318
|
|
|
seconds_delta) |
319
|
|
|
|
320
|
|
|
# Return a random ID |
321
|
|
|
return random.choice(eligible_ids) |
322
|
|
|
|
323
|
|
|
|
324
|
|
|
def load_words_from_csv(csv_file, search_term, seconds_delta=None): |
325
|
|
|
"""Load the CSV and return the top words for a given time period""" |
326
|
|
|
|
327
|
|
|
eligible_ids, matched_words = words_and_ids_from_csv(csv_file, search_term, |
328
|
|
|
seconds_delta) |
329
|
|
|
|
330
|
|
|
import most_frequent_words |
331
|
|
|
# Max tweet length is 140 |
332
|
|
|
# Let's naively set an upper limit of 140/3: |
333
|
|
|
# one-character word, comma and space |
334
|
|
|
top_words = most_frequent_words.most_frequent_words(matched_words, 140/3) |
335
|
|
|
return top_words |
336
|
|
|
|
337
|
|
|
|
338
|
|
|
# ================= WORDNIK ================== |
339
|
|
|
|
340
|
|
|
# Wordnik: get API key at http://developer.wordnik.com/ |
341
|
|
|
WORDNIK_API_KEY = "3fd3445662c1ac873962d06094f057f39d4711730e1adc28f" |
342
|
|
|
WORDNIK_USERNAME = "hugovk" |
343
|
|
|
WORDNIK_PASSWORD = "mytopsecretwordnikpassword" |
344
|
|
|
WORDNIK_TOKEN = None |
345
|
|
|
|
346
|
|
|
wordnik_client = swagger.ApiClient( |
347
|
|
|
WORDNIK_API_KEY, 'http://api.wordnik.com/v4') |
348
|
|
|
wordListApi = WordListApi.WordListApi(wordnik_client) |
349
|
|
|
|
350
|
|
|
|
351
|
|
|
# TODO: Save token to ini file |
352
|
|
|
def get_wordnik_token(): |
353
|
|
|
import getpass |
354
|
|
|
if WORDNIK_USERNAME: |
355
|
|
|
my_username = WORDNIK_USERNAME |
356
|
|
|
else: |
357
|
|
|
my_username = input("Enter your Wordnik username: ") |
358
|
|
|
if WORDNIK_PASSWORD: |
359
|
|
|
my_password = WORDNIK_PASSWORD |
360
|
|
|
else: |
361
|
|
|
my_password = getpass.getpass("Enter your Wordnik password: ") |
362
|
|
|
|
363
|
|
|
account_api = AccountApi.AccountApi(wordnik_client) |
364
|
|
|
result = account_api.authenticate(my_username, my_password) |
365
|
|
|
token = result.token |
366
|
|
|
print("Your Wordnik token is: " + token) |
367
|
|
|
return token |
368
|
|
|
|
369
|
|
|
|
370
|
|
|
def add_to_wordnik(words, wordlist_permalink): |
371
|
|
|
if len(words) == 0: |
372
|
|
|
return |
373
|
|
|
|
374
|
|
|
if TEST_MODE: |
375
|
|
|
return |
376
|
|
|
|
377
|
|
|
global WORDNIK_TOKEN |
378
|
|
|
if WORDNIK_TOKEN is None: |
379
|
|
|
# Only need to do this once |
380
|
|
|
WORDNIK_TOKEN = get_wordnik_token() |
381
|
|
|
|
382
|
|
|
words.sort() |
383
|
|
|
print_it("Words: " + ', '.join(words)) |
384
|
|
|
if len(words) == 1: |
385
|
|
|
number = "1 word" |
386
|
|
|
else: |
387
|
|
|
number = str(len(words)) + " words" |
388
|
|
|
print("Adding " + number + " to Wordnik list:" + wordlist_permalink) |
389
|
|
|
|
390
|
|
|
from wordnik.models import StringValue |
391
|
|
|
words_to_add = [] |
392
|
|
|
for word in words: |
393
|
|
|
word_to_add = StringValue.StringValue() |
394
|
|
|
word_to_add.word = word |
395
|
|
|
words_to_add.append(word_to_add) |
396
|
|
|
|
397
|
|
|
print_it(wordlist_permalink + " " + WORDNIK_TOKEN + " " + " ".join(words)) |
398
|
|
|
|
399
|
|
|
wordListApi.addWordsToWordList( |
400
|
|
|
wordlist_permalink, WORDNIK_TOKEN, body=words_to_add) |
401
|
|
|
|
402
|
|
|
print(number + " added") |
403
|
|
|
|
404
|
|
|
|
405
|
|
|
# ================ TWITTER ================== |
406
|
|
|
|
407
|
|
|
t = None |
408
|
|
|
|
409
|
|
|
|
410
|
|
|
def init_twitter(oauth_token, oauth_secret, consumer_key, consumer_secret): |
411
|
|
|
global t |
412
|
|
|
t = Twitter(auth=OAuth(oauth_token, oauth_secret, |
413
|
|
|
consumer_key, consumer_secret)) |
414
|
|
|
|
415
|
|
|
|
416
|
|
|
def get_words_from_twitter(search_term, since_id=0): |
417
|
|
|
results = t.search.tweets( |
418
|
|
|
q='"' + search_term + '"', count=100, since_id=int(since_id)) |
419
|
|
|
|
420
|
|
|
print(results['search_metadata']) |
421
|
|
|
print("Requested:\t" + str(results['search_metadata']['count'])) |
422
|
|
|
print("Found:\t" + str(len(results['statuses']))) |
423
|
|
|
max_id = results['search_metadata']['max_id'] |
424
|
|
|
print("Max ID:\t" + str(max_id)) |
425
|
|
|
|
426
|
|
|
return max_id, results |
427
|
|
|
|
428
|
|
|
|
429
|
|
|
def retweet(id, trim_user=True): |
430
|
|
|
print_it("RETWEET THIS: " + str(id)) |
431
|
|
|
|
432
|
|
|
if not TEST_MODE: |
433
|
|
|
try: |
434
|
|
|
t.statuses.retweet(id=id, trim_user=trim_user) |
435
|
|
|
except Exception as e: |
436
|
|
|
print(str(e)) |
437
|
|
|
# TODO If the account is now protected, we get an error like... |
438
|
|
|
# Twitter sent status 403 for URL: 1.1/statuses/retweet/ |
439
|
|
|
# 012345678901234567.json using parameters: ... |
440
|
|
|
# details: {"errors":"sharing is not permissible for this status |
441
|
|
|
# (Share validations failed)"} |
442
|
|
|
# ... so could try another. |
443
|
|
|
|
444
|
|
|
|
445
|
|
|
def tweet_string(string): |
446
|
|
|
if len(string) <= 0: |
447
|
|
|
return |
448
|
|
|
if len(string) + 1 <= 140: # Finish properly, if there's room |
449
|
|
|
string += "." |
450
|
|
|
|
451
|
|
|
print_it("TWEET THIS: " + string) |
452
|
|
|
|
453
|
|
|
if not TEST_MODE: |
454
|
|
|
try: |
455
|
|
|
t.statuses.update(status=string) |
456
|
|
|
except Exception as e: |
457
|
|
|
print(str(e)) |
458
|
|
|
|
459
|
|
|
|
460
|
|
|
def update_tweet_with_words(tweet, words): |
461
|
|
|
""" |
462
|
|
|
IN: tweet with a prefix, list of words |
463
|
|
|
OUT: updated tweet, list of words_remaining |
464
|
|
|
""" |
465
|
|
|
new_tweet = tweet |
466
|
|
|
words_remaining = list(words) |
467
|
|
|
for i, word in enumerate(words): |
468
|
|
|
if i == 0: |
469
|
|
|
new_tweet = tweet + word |
470
|
|
|
else: |
471
|
|
|
# new_tweet = tweet + ", " + word |
472
|
|
|
new_tweet = tweet + " " + word |
473
|
|
|
if len(new_tweet) > 140: |
474
|
|
|
break |
475
|
|
|
else: |
476
|
|
|
tweet = new_tweet |
477
|
|
|
words_remaining.pop(0) |
478
|
|
|
return tweet, words_remaining |
479
|
|
|
|
480
|
|
|
|
481
|
|
|
def tweet_those( |
482
|
|
|
words, tweet_prefix, csv_file=None, search_term=None, mode="latest"): |
483
|
|
|
# Remove duplicates |
484
|
|
|
words = dedupe(words) |
485
|
|
|
|
486
|
|
|
shuffle, tweet_all_words = False, False |
487
|
|
|
extra_prefix = "" |
488
|
|
|
|
489
|
|
|
if mode == "retweet": |
490
|
|
|
id = pick_a_random_tweet(csv_file, search_term, 2 * DAY_IN_SECONDS) |
491
|
|
|
retweet(id) |
492
|
|
|
return |
493
|
|
|
elif mode == "none": |
494
|
|
|
return |
495
|
|
|
elif mode == "latest": |
496
|
|
|
tweet_all_words = True |
497
|
|
|
elif mode == "latest_onetweet": |
498
|
|
|
shuffle = True |
499
|
|
|
elif mode == "24hours": |
500
|
|
|
words = load_words_from_csv(csv_file, search_term, DAY_IN_SECONDS) |
501
|
|
|
extra_prefix += " (24 hours)" |
502
|
|
|
elif mode == "7days": |
503
|
|
|
words = load_words_from_csv(csv_file, search_term, 7 * DAY_IN_SECONDS) |
504
|
|
|
extra_prefix += " (7 days)" |
505
|
|
|
elif mode == "30days": |
506
|
|
|
words = load_words_from_csv(csv_file, search_term, 30 * DAY_IN_SECONDS) |
507
|
|
|
extra_prefix += " (30 days)" |
508
|
|
|
elif mode == "thisyear": |
509
|
|
|
# How many seconds since 1 Jan this year? |
510
|
|
|
now = datetime.datetime.now() |
511
|
|
|
year_start = datetime.datetime(now.year, month=1, day=1) |
512
|
|
|
seconds_delta = (now - year_start).total_seconds() |
513
|
|
|
words = load_words_from_csv(csv_file, search_term, seconds_delta) |
514
|
|
|
extra_prefix += " (" + str(now.year) + ")" |
515
|
|
|
elif mode == "alltime": |
516
|
|
|
words = load_words_from_csv(csv_file, search_term, None) |
517
|
|
|
extra_prefix += " (all time)" |
518
|
|
|
else: |
519
|
|
|
print("Unknown mode: " + mode) |
520
|
|
|
return |
521
|
|
|
|
522
|
|
|
if len(words) < 1: # validation |
523
|
|
|
return |
524
|
|
|
|
525
|
|
|
if shuffle: |
526
|
|
|
random.shuffle(words) |
527
|
|
|
|
528
|
|
|
tweet = tweet_prefix |
529
|
|
|
if len(words) == 1: # get the plural right |
530
|
|
|
tweet += extra_prefix + ": " |
531
|
|
|
else: |
532
|
|
|
tweet += "s" + extra_prefix + ": " |
533
|
|
|
|
534
|
|
|
tweet, words_remaining = update_tweet_with_words(tweet, words) |
535
|
|
|
|
536
|
|
|
tweet_string(tweet) |
537
|
|
|
|
538
|
|
|
if tweet_all_words and len(words_remaining) > 0: |
539
|
|
|
tweet_those(words_remaining, tweet_prefix) |
540
|
|
|
|
541
|
|
|
# End of file |
542
|
|
|
|