Compare commits
27 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
366cfc0795 | ||
|
|
cbd4b924b8 | ||
|
|
56759e3f3c | ||
|
|
c193428b38 | ||
|
|
a625d9e9c5 | ||
|
|
36e80a73f7 | ||
|
|
cdae1cf226 | ||
|
|
bca569318e | ||
|
|
d11cd85a66 | ||
|
|
2d32a6e204 | ||
|
|
bad6456a44 | ||
|
|
1687ae66ab | ||
|
|
ddc34348d9 | ||
|
|
1d74cfeb19 | ||
|
|
1589d07b56 | ||
|
|
d261237320 | ||
|
|
66af3080dd | ||
|
|
9d396b9559 | ||
|
|
23b6ad12c1 | ||
|
|
22131e9fc7 | ||
|
|
e99e61f95a | ||
|
|
a3fe95ea27 | ||
|
|
000cb70bcb | ||
|
|
c8d9d06e75 | ||
|
|
a5e07a0375 | ||
|
|
a0a12bcf4c | ||
|
|
42e5751705 |
@@ -1,6 +1,31 @@
|
||||
Change Log
|
||||
===========
|
||||
|
||||
0.1.87
|
||||
------
|
||||
- Fix localizing midnight when non-existent (DST) #1176
|
||||
- Fix thread deadlock in bpython #1163
|
||||
|
||||
0.1.86
|
||||
------
|
||||
- Fix 'trailingPegRatio' #1141
|
||||
- Improve handling delisted tickers #1142
|
||||
- Fix corrupt tkr-tz-csv halting code #1162
|
||||
- Change default start to 1900-01-01 #1170
|
||||
|
||||
0.1.85
|
||||
------
|
||||
- Fix info['log_url'] #1062
|
||||
- Fix handling delisted ticker #1137
|
||||
|
||||
0.1.84
|
||||
------
|
||||
- Make tz-cache thread-safe
|
||||
|
||||
0.1.83
|
||||
------
|
||||
- Reduce spam-effect of tz-fetch
|
||||
|
||||
0.1.81
|
||||
------
|
||||
- Fix unhandled tz-cache exception #1107
|
||||
|
||||
10
README.md
10
README.md
@@ -274,12 +274,10 @@ To install `yfinance` using `conda`, see
|
||||
### Requirements
|
||||
|
||||
- [Python](https://www.python.org) \>= 2.7, 3.4+
|
||||
- [Pandas](https://github.com/pydata/pandas) (tested to work with
|
||||
\>=0.23.1)
|
||||
- [Numpy](http://www.numpy.org) \>= 1.11.1
|
||||
- [requests](http://docs.python-requests.org/en/master/) \>= 2.14.2
|
||||
- [lxml](https://pypi.org/project/lxml/) \>= 4.5.1
|
||||
- [appdirs](https://pypi.org/project/appdirs) \>=1.4.4
|
||||
- [Pandas](https://github.com/pydata/pandas) \>= 1.3.0
|
||||
- [Numpy](http://www.numpy.org) \>= 1.16.5
|
||||
- [requests](http://docs.python-requests.org/en/master/) \>= 2.26
|
||||
- [appdirs](https://pypi.org/project/appdirs) \>= 1.4.4
|
||||
|
||||
### Optional (if you want to use `pandas_datareader`)
|
||||
|
||||
|
||||
12
meta.yaml
12
meta.yaml
@@ -1,5 +1,5 @@
|
||||
{% set name = "yfinance" %}
|
||||
{% set version = "0.1.58" %}
|
||||
{% set version = "0.1.87" %}
|
||||
|
||||
package:
|
||||
name: "{{ name|lower }}"
|
||||
@@ -16,21 +16,19 @@ build:
|
||||
|
||||
requirements:
|
||||
host:
|
||||
- pandas >=0.24.0
|
||||
- pandas >=1.3.0
|
||||
- numpy >=1.16.5
|
||||
- requests >=2.21
|
||||
- requests >=2.26
|
||||
- multitasking >=0.0.7
|
||||
- lxml >=4.5.1
|
||||
- appdirs >= 1.4.4
|
||||
- pip
|
||||
- python
|
||||
|
||||
run:
|
||||
- pandas >=0.24.0
|
||||
- pandas >=1.3.0
|
||||
- numpy >=1.16.5
|
||||
- requests >=2.21
|
||||
- requests >=2.26
|
||||
- multitasking >=0.0.7
|
||||
- lxml >=4.5.1
|
||||
- appdirs >= 1.4.4
|
||||
- python
|
||||
|
||||
|
||||
@@ -1,6 +1,5 @@
|
||||
pandas>=0.24.0
|
||||
pandas>=1.3.0
|
||||
numpy>=1.16.5
|
||||
requests>=2.26
|
||||
multitasking>=0.0.7
|
||||
lxml>=4.5.1
|
||||
appdirs>=1.4.4
|
||||
|
||||
4
setup.py
4
setup.py
@@ -61,9 +61,9 @@ setup(
|
||||
platforms=['any'],
|
||||
keywords='pandas, yahoo finance, pandas datareader',
|
||||
packages=find_packages(exclude=['contrib', 'docs', 'tests', 'examples']),
|
||||
install_requires=['pandas>=0.24.0', 'numpy>=1.15',
|
||||
install_requires=['pandas>=1.3.0', 'numpy>=1.16.5',
|
||||
'requests>=2.26', 'multitasking>=0.0.7',
|
||||
'lxml>=4.5.1', 'appdirs>=1.4.4'],
|
||||
'appdirs>=1.4.4'],
|
||||
entry_points={
|
||||
'console_scripts': [
|
||||
'sample=sample:main',
|
||||
|
||||
@@ -15,21 +15,83 @@ Sanity check for most common library uses all working
|
||||
|
||||
import yfinance as yf
|
||||
import unittest
|
||||
import datetime
|
||||
|
||||
session = None
|
||||
import requests_cache ; session = requests_cache.CachedSession("yfinance.cache", expire_after=24*60*60)
|
||||
|
||||
symbols = ['MSFT', 'IWO', 'VFINX', '^GSPC', 'BTC-USD']
|
||||
tickers = [yf.Ticker(symbol) for symbol in symbols]
|
||||
tickers = [yf.Ticker(symbol, session=session) for symbol in symbols]
|
||||
delisted_symbols = ["BRK.B", "SDLP"]
|
||||
delisted_tickers = [yf.Ticker(symbol, session=session) for symbol in delisted_symbols]
|
||||
|
||||
|
||||
class TestTicker(unittest.TestCase):
|
||||
def setUp(self):
|
||||
d_today = datetime.date.today()
|
||||
d_today -= datetime.timedelta(days=30)
|
||||
self.start_d = datetime.date(d_today.year, d_today.month, 1)
|
||||
|
||||
def test_info_history(self):
|
||||
# always should have info and history for valid symbols
|
||||
for ticker in tickers:
|
||||
# always should have info and history for valid symbols
|
||||
assert(ticker.info is not None and ticker.info != {})
|
||||
history = ticker.history(period="max")
|
||||
history = ticker.history(period="1mo")
|
||||
assert(history.empty is False and history is not None)
|
||||
histories = yf.download(symbols, period="1mo", session=session)
|
||||
assert(histories.empty is False and histories is not None)
|
||||
|
||||
for ticker in tickers:
|
||||
assert(ticker.info is not None and ticker.info != {})
|
||||
history = ticker.history(start=self.start_d)
|
||||
assert(history.empty is False and history is not None)
|
||||
histories = yf.download(symbols, start=self.start_d, session=session)
|
||||
assert(histories.empty is False and histories is not None)
|
||||
|
||||
def test_info_history_nofail(self):
|
||||
# should not throw Exception for delisted tickers, just print a message
|
||||
for ticker in delisted_tickers:
|
||||
history = ticker.history(period="1mo")
|
||||
histories = yf.download(delisted_symbols, period="1mo", session=session)
|
||||
histories = yf.download(delisted_symbols[0], period="1mo", session=session)
|
||||
histories = yf.download(delisted_symbols[1], period="1mo")#, session=session)
|
||||
for ticker in delisted_tickers:
|
||||
history = ticker.history(start=self.start_d)
|
||||
histories = yf.download(delisted_symbols, start=self.start_d, session=session)
|
||||
histories = yf.download(delisted_symbols[0], start=self.start_d, session=session)
|
||||
histories = yf.download(delisted_symbols[1], start=self.start_d, session=session)
|
||||
|
||||
def test_attributes(self):
|
||||
for ticker in tickers:
|
||||
ticker.isin
|
||||
ticker.major_holders
|
||||
ticker.institutional_holders
|
||||
ticker.mutualfund_holders
|
||||
ticker.dividends
|
||||
ticker.splits
|
||||
ticker.actions
|
||||
ticker.info
|
||||
ticker.info["trailingPegRatio"]
|
||||
ticker.calendar
|
||||
ticker.recommendations
|
||||
ticker.earnings
|
||||
ticker.quarterly_earnings
|
||||
ticker.financials
|
||||
ticker.quarterly_financials
|
||||
ticker.balance_sheet
|
||||
ticker.quarterly_balance_sheet
|
||||
ticker.cashflow
|
||||
ticker.quarterly_cashflow
|
||||
ticker.sustainability
|
||||
ticker.options
|
||||
ticker.news
|
||||
ticker.shares
|
||||
ticker.earnings_history
|
||||
ticker.earnings_dates
|
||||
|
||||
def test_attributes_nofail(self):
|
||||
# should not throw Exception for delisted tickers, just print a message
|
||||
for ticker in delisted_tickers:
|
||||
ticker.isin
|
||||
ticker.major_holders
|
||||
ticker.institutional_holders
|
||||
|
||||
193
yfinance/base.py
193
yfinance/base.py
@@ -145,22 +145,22 @@ class TickerBase():
|
||||
debug_mode = True
|
||||
if "debug" in kwargs and isinstance(kwargs["debug"], bool):
|
||||
debug_mode = kwargs["debug"]
|
||||
if "many" in kwargs and kwargs["many"]:
|
||||
# Disable prints with threads, it deadlocks/throws
|
||||
debug_mode = False
|
||||
|
||||
err_msg = "No data found for this date range, symbol may be delisted"
|
||||
|
||||
if start or period is None or period.lower() == "max":
|
||||
# Check can get TZ. Fail => probably delisted
|
||||
try:
|
||||
tz = self._get_ticker_tz()
|
||||
except KeyError as e:
|
||||
if "exchangeTimezoneName" in str(e):
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return utils.empty_df()
|
||||
else:
|
||||
raise
|
||||
tz = self._get_ticker_tz(debug_mode, proxy, timeout)
|
||||
if tz is None:
|
||||
# Every valid ticker has a timezone. Missing = problem
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return utils.empty_df()
|
||||
|
||||
if end is None:
|
||||
end = int(_time.time())
|
||||
@@ -170,7 +170,8 @@ class TickerBase():
|
||||
if interval == "1m":
|
||||
start = end - 604800 # Subtract 7 days
|
||||
else:
|
||||
start = -631159200
|
||||
#time stamp of 01/01/1900
|
||||
start = -2208994789
|
||||
else:
|
||||
start = utils._parse_user_dt(start, tz)
|
||||
params = {"period1": start, "period2": end}
|
||||
@@ -218,7 +219,7 @@ class TickerBase():
|
||||
if data is None or not type(data) is dict or 'status_code' in data.keys():
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return utils.empty_df()
|
||||
|
||||
@@ -226,7 +227,7 @@ class TickerBase():
|
||||
err_msg = data["chart"]["error"]["description"]
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return shared._DFS[self.ticker]
|
||||
|
||||
@@ -234,7 +235,7 @@ class TickerBase():
|
||||
not data["chart"]["result"]:
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return shared._DFS[self.ticker]
|
||||
|
||||
@@ -249,7 +250,7 @@ class TickerBase():
|
||||
except Exception:
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return shared._DFS[self.ticker]
|
||||
|
||||
@@ -285,7 +286,7 @@ class TickerBase():
|
||||
err_msg = "back_adjust failed with %s" % e
|
||||
shared._DFS[self.ticker] = utils.empty_df()
|
||||
shared._ERRORS[self.ticker] = err_msg
|
||||
if "many" not in kwargs and debug_mode:
|
||||
if debug_mode:
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
|
||||
if rounding:
|
||||
@@ -315,7 +316,7 @@ class TickerBase():
|
||||
else:
|
||||
# If a midnight is during DST transition hour when clocks roll back,
|
||||
# meaning clock hits midnight twice, then use the 2nd (ambiguous=True)
|
||||
df.index = _pd.to_datetime(df.index.date).tz_localize(tz_exchange, ambiguous=True)
|
||||
df.index = _pd.to_datetime(df.index.date).tz_localize(tz_exchange, ambiguous=True, nonexistent='shift_forward')
|
||||
df.index.name = "Date"
|
||||
|
||||
# duplicates and missing rows cleanup
|
||||
@@ -331,23 +332,79 @@ class TickerBase():
|
||||
|
||||
# ------------------------
|
||||
|
||||
def _get_ticker_tz(self):
|
||||
def _get_ticker_tz(self, debug_mode, proxy, timeout):
|
||||
if not self._tz is None:
|
||||
return self._tz
|
||||
|
||||
tkr_tz = utils.cache_lookup_tkr_tz(self.ticker)
|
||||
|
||||
if tkr_tz is not None:
|
||||
invalid_value = isinstance(tkr_tz, str)
|
||||
if not invalid_value:
|
||||
try:
|
||||
_tz.timezone(tz)
|
||||
except:
|
||||
invalid_value = True
|
||||
if invalid_value:
|
||||
# Clear from cache and force re-fetch
|
||||
utils.cache_store_tkr_tz(self.ticker, None)
|
||||
tkr_tz = None
|
||||
|
||||
if tkr_tz is None:
|
||||
tkr_tz = self.info["exchangeTimezoneName"]
|
||||
# info fetch is relatively slow so cache timezone
|
||||
try:
|
||||
utils.cache_store_tkr_tz(self.ticker, tkr_tz)
|
||||
except PermissionError:
|
||||
# System probably read-only, so cannot cache
|
||||
pass
|
||||
tkr_tz = self._fetch_ticker_tz(debug_mode, proxy, timeout)
|
||||
|
||||
if tkr_tz is not None:
|
||||
try:
|
||||
utils.cache_store_tkr_tz(self.ticker, tkr_tz)
|
||||
except PermissionError:
|
||||
# System probably read-only, so cannot cache
|
||||
pass
|
||||
|
||||
self._tz = tkr_tz
|
||||
return tkr_tz
|
||||
|
||||
|
||||
def _fetch_ticker_tz(self, debug_mode, proxy, timeout):
|
||||
# Query Yahoo for basic price data just to get returned timezone
|
||||
|
||||
params = {"range":"1d", "interval":"1d"}
|
||||
|
||||
# setup proxy in requests format
|
||||
if proxy is not None:
|
||||
if isinstance(proxy, dict) and "https" in proxy:
|
||||
proxy = proxy["https"]
|
||||
proxy = {"https": proxy}
|
||||
|
||||
# Getting data from json
|
||||
url = "{}/v8/finance/chart/{}".format(self._base_url, self.ticker)
|
||||
|
||||
session = self.session or _requests
|
||||
try:
|
||||
data = session.get(url=url, params=params, proxies=proxy, headers=utils.user_agent_headers, timeout=timeout)
|
||||
data = data.json()
|
||||
except Exception as e:
|
||||
if debug_mode:
|
||||
print("Failed to get ticker '{}' reason: {}".format(self.ticker, e))
|
||||
return None
|
||||
else:
|
||||
error = data.get('chart', {}).get('error', None)
|
||||
if error:
|
||||
# explicit error from yahoo API
|
||||
if debug_mode:
|
||||
print("Got error from yahoo api for ticker {}, Error: {}".format(self.ticker, error))
|
||||
else:
|
||||
try:
|
||||
return data["chart"]["result"][0]["meta"]["exchangeTimezoneName"]
|
||||
except Exception as err:
|
||||
if debug_mode:
|
||||
print("Could not get exchangeTimezoneName for ticker '{}' reason: {}".format(self.ticker, err))
|
||||
print("Got response: ")
|
||||
print("-------------")
|
||||
print(" {}".format(data))
|
||||
print("-------------")
|
||||
return None
|
||||
|
||||
|
||||
def _get_info(self, proxy=None):
|
||||
# setup proxy in requests format
|
||||
if proxy is not None:
|
||||
@@ -425,9 +482,12 @@ class TickerBase():
|
||||
|
||||
self._info['logo_url'] = ""
|
||||
try:
|
||||
domain = self._info['website'].split(
|
||||
'://')[1].split('/')[0].replace('www.', '')
|
||||
self._info['logo_url'] = 'https://logo.clearbit.com/%s' % domain
|
||||
if not 'website' in self._info:
|
||||
self._info['logo_url'] = 'https://logo.clearbit.com/%s.com' % self._info['shortName'].split(' ')[0].split(',')[0]
|
||||
else:
|
||||
domain = self._info['website'].split(
|
||||
'://')[1].split('/')[0].replace('www.', '')
|
||||
self._info['logo_url'] = 'https://logo.clearbit.com/%s' % domain
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
@@ -457,6 +517,29 @@ class TickerBase():
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
# Complementary key-statistics. For now just want 'trailing PEG ratio'
|
||||
session = self.session or _requests
|
||||
keys = {"trailingPegRatio"}
|
||||
if len(keys)>0:
|
||||
# For just one/few variable is faster to query directly:
|
||||
url = "https://query1.finance.yahoo.com/ws/fundamentals-timeseries/v1/finance/timeseries/{}?symbol={}".format(self.ticker, self.ticker)
|
||||
for k in keys:
|
||||
url += "&type="+k
|
||||
# Request 6 months of data
|
||||
url += "&period1={}".format(int((_datetime.datetime.now()-_datetime.timedelta(days=365//2)).timestamp()))
|
||||
url += "&period2={}".format(int((_datetime.datetime.now()+_datetime.timedelta(days=1)).timestamp()))
|
||||
json_str = session.get(url=url, proxies=proxy, headers=utils.user_agent_headers).text
|
||||
json_data = _json.loads(json_str)
|
||||
key_stats = json_data["timeseries"]["result"][0]
|
||||
if k not in key_stats:
|
||||
# Yahoo website prints N/A, indicates Yahoo lacks necessary data to calculate
|
||||
v = None
|
||||
else:
|
||||
# Select most recent (last) raw value in list:
|
||||
v = key_stats[k][-1]["reportedValue"]["raw"]
|
||||
self._info[k] = v
|
||||
|
||||
|
||||
def _get_fundamentals(self, proxy=None):
|
||||
def cleanup(data):
|
||||
df = _pd.DataFrame(data).drop(columns=['maxAge'])
|
||||
@@ -616,48 +699,6 @@ class TickerBase():
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
# Complementary key-statistics (currently fetching the important trailingPegRatio which is the value shown in the website)
|
||||
res = {}
|
||||
try:
|
||||
my_headers = {'user-agent': 'curl/7.55.1', 'accept': 'application/json', 'content-type': 'application/json',
|
||||
'referer': 'https://finance.yahoo.com/', 'cache-control': 'no-cache', 'connection': 'close'}
|
||||
p = _re.compile(r'root\.App\.main = (.*);')
|
||||
r = _requests.session().get('https://finance.yahoo.com/quote/{}/key-statistics?p={}'.format(self.ticker,
|
||||
self.ticker), headers=my_headers)
|
||||
q_results = {}
|
||||
my_qs_keys = ['pegRatio'] # QuoteSummaryStore
|
||||
# , 'quarterlyPegRatio'] # QuoteTimeSeriesStore
|
||||
my_ts_keys = ['trailingPegRatio']
|
||||
|
||||
# Complementary key-statistics
|
||||
data = _json.loads(p.findall(r.text)[0])
|
||||
key_stats = data['context']['dispatcher']['stores']['QuoteTimeSeriesStore']
|
||||
q_results.setdefault(self.ticker, [])
|
||||
for i in my_ts_keys:
|
||||
# j=0
|
||||
try:
|
||||
# res = {i: key_stats['timeSeries'][i][1]['reportedValue']['raw']}
|
||||
# We need to loop over multiple items, if they exist: 0,1,2,..
|
||||
zzz = key_stats['timeSeries'][i]
|
||||
for j in range(len(zzz)):
|
||||
if key_stats['timeSeries'][i][j]:
|
||||
res = {i: key_stats['timeSeries']
|
||||
[i][j]['reportedValue']['raw']}
|
||||
q_results[self.ticker].append(res)
|
||||
|
||||
# print(res)
|
||||
# q_results[ticker].append(res)
|
||||
except:
|
||||
q_results[ticker].append({i: np.nan})
|
||||
|
||||
res = {'Company': ticker}
|
||||
q_results[ticker].append(res)
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
if 'trailingPegRatio' in res:
|
||||
self._info['trailingPegRatio'] = res['trailingPegRatio']
|
||||
|
||||
self._fundamentals = True
|
||||
|
||||
def get_recommendations(self, proxy=None, as_dict=False, *args, **kwargs):
|
||||
@@ -803,6 +844,10 @@ class TickerBase():
|
||||
self.get_info(proxy=proxy)
|
||||
if "shortName" in self._info:
|
||||
q = self._info['shortName']
|
||||
if q is None:
|
||||
err_msg = "Cannot map to ISIN code, symbol may be delisted"
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return None
|
||||
|
||||
url = 'https://markets.businessinsider.com/ajax/' \
|
||||
'SearchController_Suggest?max_results=25&query=%s' \
|
||||
@@ -901,8 +946,10 @@ class TickerBase():
|
||||
dates = _pd.concat([dates, data], axis=0)
|
||||
page_offset += page_size
|
||||
|
||||
if dates is None:
|
||||
raise Exception("No data found, symbol may be delisted")
|
||||
if (dates is None) or dates.shape[0]==0:
|
||||
err_msg = "No earnings dates found, symbol may be delisted"
|
||||
print('- %s: %s' % (self.ticker, err_msg))
|
||||
return None
|
||||
dates = dates.reset_index(drop=True)
|
||||
|
||||
# Drop redundant columns
|
||||
|
||||
@@ -198,7 +198,7 @@ def _download_one_threaded(ticker, start=None, end=None,
|
||||
|
||||
data = _download_one(ticker, start, end, auto_adjust, back_adjust,
|
||||
actions, period, interval, prepost, proxy, rounding,
|
||||
keepna, timeout)
|
||||
keepna, timeout, many=True)
|
||||
shared._DFS[ticker.upper()] = data
|
||||
if progress:
|
||||
shared._PROGRESS_BAR.animate()
|
||||
@@ -208,11 +208,11 @@ def _download_one(ticker, start=None, end=None,
|
||||
auto_adjust=False, back_adjust=False,
|
||||
actions=False, period="max", interval="1d",
|
||||
prepost=False, proxy=None, rounding=False,
|
||||
keepna=False, timeout=None):
|
||||
keepna=False, timeout=None, many=False):
|
||||
|
||||
return Ticker(ticker).history(period=period, interval=interval,
|
||||
start=start, end=end, prepost=prepost,
|
||||
actions=actions, auto_adjust=auto_adjust,
|
||||
back_adjust=back_adjust, proxy=proxy,
|
||||
rounding=rounding, keepna=keepna, many=True,
|
||||
timeout=timeout)
|
||||
rounding=rounding, keepna=keepna, timeout=timeout,
|
||||
many=many)
|
||||
|
||||
@@ -31,6 +31,9 @@ import sys as _sys
|
||||
import os as _os
|
||||
import appdirs as _ad
|
||||
|
||||
from threading import Lock
|
||||
mutex = Lock()
|
||||
|
||||
try:
|
||||
import ujson as _json
|
||||
except ImportError:
|
||||
@@ -332,27 +335,36 @@ def cache_lookup_tkr_tz(tkr):
|
||||
if not _os.path.isfile(fp):
|
||||
return None
|
||||
|
||||
df = _pd.read_csv(fp)
|
||||
f = df["Ticker"] == tkr
|
||||
if sum(f) == 0:
|
||||
mutex.acquire()
|
||||
df = _pd.read_csv(fp, index_col="Ticker", on_bad_lines="skip")
|
||||
mutex.release()
|
||||
if tkr in df.index:
|
||||
return df.loc[tkr,"Tz"]
|
||||
else:
|
||||
return None
|
||||
|
||||
return df["Tz"][f].iloc[0]
|
||||
def cache_store_tkr_tz(tkr,tz):
|
||||
df = _pd.DataFrame({"Ticker":[tkr], "Tz":[tz]})
|
||||
|
||||
dp = get_cache_dirpath()
|
||||
fp = _os.path.join(dp, "tkr-tz.csv")
|
||||
mutex.acquire()
|
||||
if not _os.path.isdir(dp):
|
||||
_os.makedirs(dp)
|
||||
fp = _os.path.join(dp, "tkr-tz.csv")
|
||||
if not _os.path.isfile(fp):
|
||||
df.to_csv(fp, index=False)
|
||||
return
|
||||
if (not _os.path.isfile(fp)) and (tz is not None):
|
||||
df = _pd.DataFrame({"Tz":[tz]}, index=[tkr])
|
||||
df.index.name = "Ticker"
|
||||
df.to_csv(fp)
|
||||
|
||||
df_all = _pd.read_csv(fp)
|
||||
f = df_all["Ticker"]==tkr
|
||||
if sum(f) > 0:
|
||||
raise Exception("Tkr {} tz already in cache".format(tkr))
|
||||
|
||||
_pd.concat([df_all,df]).to_csv(fp, index=False)
|
||||
else:
|
||||
df = _pd.read_csv(fp, index_col="Ticker", on_bad_lines="skip")
|
||||
if tz is None:
|
||||
# Delete if in cache:
|
||||
if tkr in df.index:
|
||||
df.drop(tkr).to_csv(fp)
|
||||
else:
|
||||
if tkr in df.index:
|
||||
raise Exception("Tkr {} tz already in cache".format(tkr))
|
||||
df.loc[tkr,"Tz"] = tz
|
||||
df.to_csv(fp)
|
||||
|
||||
mutex.release()
|
||||
|
||||
|
||||
@@ -1 +1 @@
|
||||
version = "0.1.81"
|
||||
version = "0.1.87"
|
||||
|
||||
Reference in New Issue
Block a user