1
- import random
2
- import shutil
1
+ from concurrent import futures
3
2
from datetime import datetime , date , time
4
3
import logging
5
- import time as _time
6
- from concurrent import futures
4
+ import shutil
7
5
8
- import numpy as np
6
+ from dateutil . relativedelta import relativedelta , FR
9
7
import pandas as pd
10
- import requests
11
8
from tqdm import tqdm
12
- from dateutil .relativedelta import relativedelta , FR
13
-
14
- from opensignals import utils
15
9
16
10
logger = logging .getLogger (__name__ )
17
11
@@ -168,7 +162,7 @@ def get_data(
168
162
return train_data , test_data , live_data , feature_names
169
163
170
164
171
- def download_tickers (tickers , start ):
165
+ def download_tickers (tickers , start , download_ticker ):
172
166
start_epoch = int (datetime .strptime (start , '%Y-%m-%d' ).timestamp ())
173
167
end_epoch = int (datetime .combine (date .today (), time ()).timestamp ())
174
168
@@ -195,7 +189,7 @@ def download_tickers(tickers, start):
195
189
return pd .concat (dfs )
196
190
197
191
198
- def download_data (db_dir , recreate = False ):
192
+ def download_data (db_dir , download_ticker , recreate = False ):
199
193
if recreate :
200
194
logging .warning (f'Removing dataset { db_dir } to recreate it' )
201
195
shutil .rmtree (db_dir , ignore_errors = True )
@@ -218,7 +212,7 @@ def download_data(db_dir, recreate=False):
218
212
)
219
213
concat_dfs = []
220
214
for start_date , tickers in ticker_missing_grouped .iteritems ():
221
- temp_df = download_tickers (tickers .split (' ' ), start = start_date )
215
+ temp_df = download_tickers (tickers .split (' ' ), start_date , download_ticker )
222
216
223
217
# Yahoo Finance returning previous day in some situations
224
218
# (e.g. Friday in TelAviv markets)
0 commit comments