Compare commits

...

424 Commits

Author SHA1 Message Date
ValueRaider
a1bcb4c351 Version 0.2.38 2024-04-16 22:18:37 +01:00
ValueRaider
18089f451b Merge pull request #1908 from vittoboa/holders_404_error
Fix 404 Error for Holders
2024-04-16 22:14:33 +01:00
vittoboa
3d180fcf2c Move ticker symbol from parameter to URL 2024-04-16 22:37:07 +02:00
Value Raider
c2d568367c Version 0.2.37 2024-02-25 13:25:28 +00:00
ValueRaider
d3728d3071 Merge pull request #1869 from ranaroussi/dev
Dev
2024-02-24 23:09:34 +00:00
ValueRaider
915bb1a080 Merge pull request #1866 from ranaroussi/fix/price-repair-confusing-order
Price repair bug-fix
2024-02-24 22:58:20 +00:00
Value Raider
d55c317158 Fix bug: prices order flipping during repair, introducing potential data corruption 2024-02-19 22:17:20 +00:00
ValueRaider
ac1d09049e Merge pull request #1865 from cottrell/fix
Fix some errors.
2024-02-19 22:11:48 +00:00
David Cottrell
afb4e0d5dc Fix some errors. 2024-02-19 21:43:12 +00:00
ValueRaider
1d31e7ca01 Update issue form - more emphasis on following instructions 2024-02-11 13:47:36 +00:00
ValueRaider
683064f9ad Merge pull request #1849 from ranaroussi/refactor/price-history 2024-02-07 23:11:51 +00:00
Value Raider
cdf897f9e6 Move price history+repair logic into new file 2024-02-04 13:09:37 +00:00
ValueRaider
eab6c8dfa7 Update bug_report.yaml because people can't read 2024-02-01 21:28:38 +00:00
ValueRaider
97f93d35ed Merge pull request #1844 from power-edge/dev
adding upgrade for pandas deprecation warning, adding pyarrow>=0.17.0…
2024-01-31 21:51:05 +00:00
Nikolaus Schuetz
5aef8addab removing dev requirements (they are included by extras) 2024-01-29 17:43:12 -05:00
ValueRaider
6b8a4a5608 Merge pull request #1841 from Rogach/pr/dont-disable-global-logging
do not disable app-wide logging in quote.py (fixes #1829)
2024-01-28 16:29:43 +00:00
Platon Pronko
212a7987c3 do not disable app-wide logging in quote.py (fixes #1829) 2024-01-28 20:43:50 +05:00
Nikolaus Schuetz
58a0a57457 adding upgrade for pandas deprecation warning, adding pyarrow>=0.17.0 at minimum requirement as defined in dev requirements for pandas==1.3.0 version 2024-01-26 20:12:58 -05:00
ValueRaider
75297c0eba Merge pull request #1838 from mreiche/bugfix/remove-empty-series
Remove _empty_series leftovers
2024-01-23 19:07:16 +00:00
Mike Reiche
1dc2719368 Remove _empty_series leftovers 2024-01-23 15:32:56 +01:00
Value Raider
ab979e9141 Version 0.2.36 2024-01-21 18:10:41 +00:00
ValueRaider
b837c1ec2a Merge pull request #1834 from ranaroussi/dev
sync dev -> main
2024-01-21 18:08:04 +00:00
ValueRaider
2630c66cd1 Merge pull request #1833 from ange-daumal/json-fix
Fix JSON error handling
2024-01-19 21:56:42 +00:00
ValueRaider
7af789fe9a Merge pull request #1830 from ange-daumal/patch-1
Fix JSON error handling
2024-01-19 21:51:52 +00:00
ValueRaider
73e36688b7 Merge pull request #1827 from ranaroussi/fix/peewee-with-old-sqlite
Handle peewee with old sqlite
2024-01-19 21:51:31 +00:00
ValueRaider
f1264716fc Merge pull request #1824 from ranaroussi/fix/price-keepna-with-repair
Fix history() keepna=False with repair=True
2024-01-19 21:51:10 +00:00
Ange Daumal
06fd35121a Fix JSON access to prevent KeyError 2024-01-19 22:51:02 +01:00
Mike Reiche
91f468e4d3 Fix JSON access to prevent KeyError 2024-01-19 22:42:42 +01:00
ValueRaider
d00c1a976c Merge pull request #1831 from ranaroussi/main
sync main -> dev
2024-01-15 19:28:56 +00:00
ValueRaider
176c3d628b Update ci.yml to Node16 2024-01-15 19:27:37 +00:00
ValueRaider
8f53af1593 Merge pull request #1823 from molpcs/patch-2
Update README.md for better copy-ability
2024-01-14 12:36:17 +00:00
ValueRaider
19188d52d4 Merge pull request #1795 from amanlai/main
explicitly name the column levels
2024-01-14 10:45:04 +00:00
Value Raider
ffaf200562 Handle peewee with old sqlite 2024-01-13 23:00:59 +00:00
Value Raider
6686258e66 Fix history() keepna=False with repair=True 2024-01-13 13:19:44 +00:00
molpcs
47bc46c804 Update README.md
Wrap yfinance[optional] code snippet with quotes to avoid conflict with zsh globbing. Remains compatible with bash.
2024-01-12 11:57:58 -08:00
ValueRaider
f563e51509 Merge pull request #1822 from akshayparopkari/patch-1 2024-01-11 09:30:22 +00:00
Akshay Paropkari
c5404bcd9d Update fundamentals.py
Error in supplying timescale values resulted in misleading ValueError - 

```
ValueError: Illegal argument: timescale must be one of: ['income', 'balance-sheet', 'cash-flow']
```
2024-01-11 05:17:39 +00:00
ValueRaider
006e0a155b Merge pull request #1724 from mreiche/bugfix/data-types-2
Bugfix/data types 2
2024-01-09 20:13:41 +00:00
Mike Reiche
dbc55e5596 Remove unused List import 2024-01-09 21:08:46 +01:00
Mike Reiche
4ce63fe8ca Merge remote-tracking branch 'yfinance/dev' into bugfix/data-types-2 2024-01-09 08:51:33 +01:00
Mike Reiche
223f5337a8 Remove empty static series 2024-01-09 08:50:31 +01:00
Mike Reiche
4c34487149 Revert disabling earnings test 2024-01-09 08:50:00 +01:00
Mike Reiche
ac8a917288 Revert adding explicit requirements 2024-01-09 08:43:54 +01:00
Mike Reiche
15321bd097 Merge remote-tracking branch 'yfinance/main' into bugfix/data-types-2 2024-01-09 08:42:43 +01:00
ValueRaider
10961905b6 Merge pull request #1817 from ranaroussi/main
sync main -> dev
2024-01-07 18:39:10 +00:00
ValueRaider
acbd2a8d78 Merge pull request #1816 from ranaroussi/fix/ticker-api
0.2.34 fixes: Add new data to README, remove deprecated stuff, fix tests
2024-01-07 00:35:06 +00:00
Value Raider
61c4696c65 Add new data to README, remove deprecated stuff, fix tests, v0.2.35
Ticker.recommendations*:
- add to README
- organise their unit tests
- remove redundant recommendations_history

Remove deprecated arguments from Ticker.history

Fix 'bad symbol' behaviour & tests
Fix some prices tests

Bump version 0.2.35
2024-01-07 00:33:59 +00:00
Value Raider
a7c41afa52 Version 0.2.34 2024-01-06 17:19:51 +00:00
ValueRaider
49d8dfd544 Merge pull request #1815 from ranaroussi/dev
sync dev -> main
2024-01-06 16:18:20 +00:00
ValueRaider
477dc6e6c4 Merge pull request #1798 from ranaroussi/fix/price-repair-div-adjust
Fix price repair div adjust
2023-12-31 21:43:44 +00:00
ValueRaider
7e6ad0834c Merge pull request #1806 from puntonim/ticker-history-exc-hook
Ticker.history() to raise HTTP request excs if raise_errors args is True
2023-12-31 14:09:12 +00:00
puntonim
c94cbb64d4 Ticker.history() to raise HTTP request excs if raise_errors args is True 2023-12-31 14:57:47 +01:00
ValueRaider
c053e2cb30 Merge pull request #1807 from ranaroussi/feature/optional-reqs-min-versions
Set sensible min versions for optional 'nospam' reqs
2023-12-31 13:47:17 +00:00
Value Raider
112b297c41 Set sensible min versions for optional 'nospam' reqs
Set sensible min versions for optional 'nospam' reqs:
- requests_cache >= 1.0 , first defined DO_NOT_CACHE
2023-12-31 13:45:26 +00:00
ValueRaider
5195c3a798 Merge pull request #1810 from Tejasweee/dev
make nan as float
2023-12-31 12:56:26 +00:00
Tejasweee
c1ad2589da make nan as float 2023-12-31 09:29:19 +05:45
ValueRaider
d1a34a4da0 Merge pull request #1796 from ranaroussi/fix/cookie-cache-date
Fix invalid date entering cache DB
2023-12-30 17:32:14 +00:00
Value Raider
d44eff4065 Fix 'Unalignable' error in reconstruct_intervals 2023-12-22 20:29:04 +00:00
Value Raider
db670aefd7 Fix invalid date entering cache DB
'peewee.DateTimeField' is not ISO-compliant. If user enforces strict ISO-compliance,
then translation between DateTimeField and sqlite breaks. Fix is to manually
implement translation.
2023-12-22 12:59:50 +00:00
Manlai Amar
a3095d2a40 explicitly name the column levels 2023-12-21 00:02:53 -08:00
ValueRaider
f753e6090d Merge pull request #1793 from ranaroussi/fix/fetch-tkr-tz
Fix _get_ticker_tz() args, were being swapped. Improve its unit test
2023-12-17 18:59:06 +00:00
Value Raider
9021fe52b4 Fix _get_ticker_tz() args, were being swapped. Improve its unit test 2023-12-17 18:35:29 +00:00
ValueRaider
281cc64a4a Merge pull request #1790 from bot-unit/feature/calendar
feature calendar events
2023-12-16 13:37:19 +00:00
ValueRaider
8975689bd1 README: add cache folder location 2023-12-16 13:36:40 +00:00
Unit
24f53e935d added calendar events
added events from calendarEvents module
returning data is dict
test upgraded and passed
2023-12-16 13:35:04 +01:00
ValueRaider
a6790606ef Merge pull request #1774 from coskos-ops/fix/complementaryinfo
Fixed incorrect code for ticker complementary info retrieval
2023-12-14 17:57:26 +00:00
Filip Kostic
122269cf53 Fixed fstring error 2023-12-13 19:45:47 -05:00
ValueRaider
a914647fa4 Merge pull request #1772 from JuliaLWang8/feat/holders-insiders
Feat/Holders insider data
2023-12-13 22:13:24 +00:00
Julia L. Wang
dc957eeb0e Implementation of holders data 2023-12-13 16:57:13 -05:00
ValueRaider
f8d65d0def Merge pull request #1773 from bot-unit/feature/upgrades_downgrades
add upgrades downgrades
2023-12-13 20:59:58 +00:00
ValueRaider
f32097e157 Merge pull request #1771 from JuliaLWang8/feat/extra-dependencies
Feat/adding extra dependencies
2023-12-13 20:58:48 +00:00
Value Raider
469037be80 Tweaks to formatting and links. 2023-12-13 19:26:05 +00:00
Julia L. Wang
9648e69b7e Updated scipy and readme 2023-12-12 18:26:04 -05:00
ValueRaider
f718db6c2f Merge pull request #1776 from coskos-ops/fix/progressSTDerr 2023-12-12 23:08:28 +00:00
Filip Kostic
c8280e4001 Update utils.py 2023-12-12 17:45:26 -05:00
ValueRaider
53c29480b6 Merge pull request #1779 from VishnuAkundi/invalid_symbol_date_fix
Fix for Key Error Issue on Date column when one of the symbols is no longer valid (delisted/not available)
2023-12-12 21:16:39 +00:00
Vishnu Akundi
4a5616d5c4 Added Fix and Unit Test for Issue 2023-12-12 11:29:58 -05:00
Filip Kostic
5e0006e4b3 Removed redundant import 2023-12-11 15:07:16 -05:00
Filip Kostic
2b1a26ef0c Moved progress bar output to stderr 2023-12-10 20:51:11 -05:00
Filip Kostic
8fdf53233f Fixed issue #1305. Added test case to test for trailingPegInfo statistic retrieval 2023-12-10 17:54:08 -05:00
Unit
4175885747 add upgrades downgrades
add upgrades/downgrades (recommendations history)
return data is pandas dataframe
add test for upgrades/downgrades data
2023-12-10 22:35:53 +01:00
ValueRaider
580502941a Merge pull request #1766 from JuliaLWang8/pandas-future-proofing
Pandas future proofing
2023-12-10 20:33:46 +00:00
Julia L. Wang
1863b211cd Added extra dependencies 2023-12-10 10:36:02 -05:00
Julia L. Wang
0bcd2dc725 Removed unnecessary iloc 2023-12-09 23:08:22 -05:00
ValueRaider
c60e590bd7 Merge pull request #1768 from ranaroussi/fix/price-repair-and-tests
Minor fixes for price repair and related tests
2023-12-09 21:45:25 +00:00
ValueRaider
fce4707340 Merge pull request #1756 from marco-carvalho/ruff
Add Ruff
2023-12-09 21:33:28 +00:00
Value Raider
f7825c1c3a Minor fixes for price repair and related tests
Minor fixes for price repair and related tests:
- update out-of-date test, remove delisted ticker
- fix Numpy type mismatch error
2023-12-09 19:40:20 +00:00
Marco Carvalho
27ef2bcd1a Update ruff.yml 2023-12-09 13:18:35 +00:00
Marco Carvalho
fb2006b814 add ruff 2023-12-09 13:18:35 +00:00
Julia L. Wang
9b9158050a Pandas future proofing (tested)
Changed fillna, iloc, and added test changes
2023-12-08 04:26:04 -05:00
ValueRaider
f30e4ebd4c Merge pull request #1764 from ranaroussi/main
sync main -> dev
2023-12-07 09:40:56 +00:00
Value Raider
f08fe83290 Version 0.2.33 2023-12-06 19:49:23 +00:00
ValueRaider
ca2040f5fd Merge pull request #1759 from ranaroussi/hotfix/cookie-fallback-strategy
Fix '_set_cookie_strategy'
2023-12-06 19:47:40 +00:00
ValueRaider
1cfeddff59 Merge pull request #1760 from bot-unit/bugfixes/from_isin
fix base class init method
2023-12-06 19:47:32 +00:00
Value Raider
1ab476b14f Fix '_set_cookie_strategy', was double-toggling. + more logging 2023-12-06 19:46:01 +00:00
Unit
ae2ae7bce4 fix class init method 2023-12-05 19:19:03 +00:00
ValueRaider
1d3ef4f733 Merge pull request #1754 from bot-unit/feature/recommendations
Feature/recommendations
2023-12-02 19:43:35 +00:00
Unit
a3ac9fc72d added recommendations
added valid modules for quote summary request
added _fetch method for fetching quote summary
added fetch recommendationTrend
2023-12-02 15:46:17 +01:00
ValueRaider
03a1f03583 Create CODE_OF_CONDUCT.md 2023-11-26 13:08:54 +00:00
ValueRaider
af9a356fd5 Merge pull request #1745 from ranaroussi/main
sync main -> dev
2023-11-19 11:10:03 +00:00
Value Raider
9b6e35bdcd Version 0.2.32 2023-11-18 12:56:06 +00:00
ValueRaider
4d4e56cdc8 Merge pull request #1657 from ranaroussi/feature/cookie-and-crumb
Add cookie & crumb to requests
2023-11-18 12:54:24 +00:00
Value Raider
91efcd8f7d Final tidy before merge 2023-11-18 12:53:42 +00:00
Value Raider
63a3531edc Remove dependence on python>3.8 2023-11-16 20:29:02 +00:00
Value Raider
1b0d8357d6 Beta version 0.2.32b1 2023-11-13 20:29:51 +00:00
Value Raider
4466e57b95 Add cookie & crumb to requests
Add cookie & crumb to requests. Involves several changes:
- fetch cookie & crumb, obviously.
- two different cookie strategies - one seems to work better in USA, other better outside.
- yfinance auto-detects if one strategy fails, and switches to other strategy.
- cookie is stored in persistent cache folder, alongside timezones. Refetched after 24 hours.

To have this work well with multithreading (yfinance.download()) requires more changes:
- all threads share the same cookie, therefore the same session object. Requires thread-safety ...
- converted data class to a singleton with "SingletonMeta":
 - the first init() call initialises data.
 - but successive calls update its session object - naughty but necessary.
- thread locks to avoid deadlocks and race conditions.
2023-11-13 19:35:12 +00:00
ValueRaider
6d3d6b659c Merge pull request #1740 from mikez/main
Fix pandas FutureWarning: "Passing literal html to 'read_html' is deprecated"
2023-11-09 20:20:56 +00:00
Value Raider
b696add360 Restore 'earnings_dates' unit tests 2023-11-09 20:18:32 +00:00
Michael B.
06751a0b9c Fix pandas FutureWarning: "Passing literal html to 'read_html' is deprecated"
This addresses #1685 (`institutional_holders`) and also `get_earnings_dates()`.

Pandas issue is found here:
https://github.com/pandas-dev/pandas/issues/53767
and the change in code here:
5cedf87ccc/pandas/io/html.py (L1238)

As for legacy Python 2.7 support: `io.StringIO` seems to be supported in
the versions I tested. See https://docs.python.org/2/library/io.html
2023-11-09 17:42:22 +01:00
Mike Reiche
ba3c1b5ac6 Merge remote-tracking branch 'yfinance/dev' into bugfix/data-types-2
# Conflicts:
#	yfinance/base.py
2023-10-21 12:44:39 +02:00
ValueRaider
7432d2939c Merge pull request #1711 from rickturner2001/refactor/ticker-proxy
Refactor/ticker proxy
2023-10-18 20:08:05 +01:00
Mike Reiche
ba977a16a2 Added tests 2023-10-12 08:53:16 +02:00
Mike Reiche
9a3d60105c Minor typing fixes 2023-10-12 08:53:16 +02:00
Mike Reiche
0521428f69 Fixed typing bug when series are empty 2023-10-12 08:53:15 +02:00
Value Raider
308e58b914 Bump version to 0.2.31 2023-10-04 22:03:24 +01:00
ValueRaider
f6beadf448 Merge pull request #1716 from ranaroussi/dev
sync dev -> main
2023-10-04 22:01:43 +01:00
rickturner2001
d607c43967 refactored Ticker proxy attribute 2023-10-01 21:46:59 -04:00
rickturner2001
4c1669ad9d Refactored tests for Ticker with proxy
Ticker proxy refactor
2023-10-01 21:08:49 -04:00
Value Raider
7da64b679e Dev version 0.2.31b2 2023-10-01 22:40:15 +01:00
ValueRaider
38f8ccd40a Merge pull request #1709 from ranaroussi/feature/tz-cache-lazy-load
Feature/tz cache lazy load
2023-10-01 22:39:19 +01:00
ValueRaider
13acc3dc97 Merge pull request #1707 from rickturner2001/fix-testing
Test Fix: Check for type and expect exceptions in tests
2023-10-01 21:19:31 +01:00
Value Raider
cc1ac7bbcc Fix cache on read-only filesystem, + tests 2023-10-01 21:17:59 +01:00
rickturner2001
75449fd0ac Merge branch 'dev' into fix-testing 2023-10-01 15:08:12 -04:00
Value Raider
22e0c414c4 Rename for clarity 2023-10-01 13:04:48 +01:00
Value Raider
37d60e6efb Complete TZ cache lazy-loading
The initial singleton design pattern for database access meant that lazy-loading was broken,
due to structure of '_KV' class. So errors were blocking import.
Fix = use 'peewee' proxy database and initialise when needed.
2023-10-01 12:53:49 +01:00
Value Raider
dac9a48742 Dev version 0.2.31b1 2023-10-01 10:06:34 +01:00
rickturner2001
bd52326091 Fix testing: Fixed broken tests and refactored code 2023-09-30 18:23:03 -04:00
ValueRaider
9581b8bd45 Merge pull request #1705 from ranaroussi/fix/tz-cache-init
Fix TZ cache exception blocking import
2023-09-30 21:16:04 +01:00
Value Raider
62b2c25da8 Disable broken Ticker tests 2023-09-27 21:45:46 +01:00
Value Raider
7618dda5d0 Hopefully fix TZ cache exception blocking import
Hopefully fix TZ cache exception blocking import. Also:
- relocate cache init lock
- add test for setTzCacheLocation()
2023-09-27 21:34:59 +01:00
ValueRaider
95ef486e13 Merge pull request #1704 from ranaroussi/main
sync main -> dev
2023-09-27 20:53:42 +01:00
ValueRaider
9e59f6b61c Merge pull request #1703 from ranaroussi/fix/prices-intraday-merge-events
Fix merging pre-market events with intraday prices
2023-09-27 20:52:23 +01:00
Value Raider
716cd65fd3 Fix merging pre-market events with intraday prices 2023-09-25 22:19:24 +01:00
Value Raider
5b1605b5a1 Bump version to 0.2.30 2023-09-24 22:18:30 +01:00
ValueRaider
412cfbcd6d Merge pull request #1698 from ranaroussi/hotfix/download-database-error
Fix: OperationalError('unable to open database file')
2023-09-24 22:15:15 +01:00
Value Raider
6abee6df44 Remove unrelated change (will go in another PR) 2023-09-23 15:20:06 +01:00
Value Raider
fad21dfeac Enhance testing - use new cache folder 2023-09-23 13:41:59 +01:00
Value Raider
fc27f9c367 Refactor TZ cache 2023-09-23 13:30:58 +01:00
Value Raider
bb79b573ed Fix download() ''unable to open database file' 2023-09-23 11:40:39 +01:00
ValueRaider
127b53ee7f Merge pull request #1693 from ranaroussi/main
sync main -> dev
2023-09-22 14:52:36 +01:00
Value Raider
88525abcbd Bump version to 0.2.29 2023-09-22 12:00:47 +01:00
ValueRaider
99ef055cc4 Merge pull request #1692 from ranaroussi/dev
sync dev -> main
2023-09-22 12:00:06 +01:00
ValueRaider
0f36f7980b Merge pull request #1688 from ranaroussi/fix/price-repair
Price repair fixes
2023-09-22 11:27:38 +01:00
ValueRaider
8282af9ce4 Merge pull request #1684 from ranaroussi/fix/prices-intraday-merge-events
Fix merging events with intraday prices
2023-09-22 11:24:30 +01:00
Value Raider
5208c8cf05 Price repair improvements
Price repair improvements:
- don't attempt repair of empty prices table
- random-mixups: fix 0.01x errors, not just 100x
- stop zeroes, big-dividends, and 100x errors triggering false split errors
2023-09-22 11:21:17 +01:00
Value Raider
d3dfb4c6a8 Fix merging events with intraday prices
If Yahoo returns intraday price data with dividend or stock-split event in future, then this broke the merge.
Fix is to discard out-of-range events.
Assumes that if user requesting intraday then they aren't interested in events.
2023-09-19 19:35:03 +01:00
ValueRaider
279726afe4 Merge pull request #1687 from arduinocc04/arduinocc04-patch-2
Fix error when calling enable_debug_mode twice
2023-09-19 17:41:28 +01:00
조다니엘(Daniel Cho)
937386f3ef Fix error when calling enable_debug_mode twice 2023-09-19 10:02:28 +09:00
ValueRaider
32e569f652 Merge pull request #1675 from ranaroussi/hotfix/database-error
Replace sqlite3 with peewee for 100% thread-safety
2023-09-17 21:47:57 +01:00
ValueRaider
de59f0b2c6 Bug template - add section to describe bug 2023-09-09 18:32:32 +01:00
Value Raider
7d6d8562e8 Replace sqlite3 with peewee for 100% thread-safety 2023-09-03 16:47:36 +01:00
ValueRaider
6cae6d45b1 Merge pull request #1672 from difelice/fix-fix_yahoo_returning_live_separate-warnings
Fix pandas warning when retrieving quotes.
2023-09-01 22:07:18 +01:00
Alessandro Di Felice
ec3de0710d Fix pandas warning when retrieving quotes 2023-09-01 14:07:53 -05:00
Value Raider
0713d93867 Bump version to 0.2.28 2023-08-13 13:05:02 +01:00
Value Raider
67e81a8f9a Fix TypeError: 'FastInfo' object is not callable #1636 2023-08-13 13:01:53 +01:00
ValueRaider
b6372c0945 Merge pull request #1661 from ranaroussi/dev
sync dev -> main
2023-08-13 12:44:05 +01:00
ValueRaider
c9dd582dd8 Merge branch 'main' into dev 2023-08-13 12:40:11 +01:00
ValueRaider
677f3d5702 Merge pull request #1660 from ranaroussi/fix/price-repair-100x-and-calibration
Fix price repair: 100x & calibration
2023-08-13 12:21:55 +01:00
Value Raider
4f9b05a546 Fix price repair: 100x & calibration
Price repair fixes and improvement

Fixes:
- fix reconstruction mis-calibration with tiny DataFrames
- fix detecting last-active-trading-interval when NaNs in DataFrame
- redesign mapping 100x signals to ranges:
  - no change for signals before last-active-trading-interval
  - but for signals after last-active-trading-interval, process in reverse order

Improvements:
- increase max reconstruction depth from 1 to 2. E.g. now 1wk can be repaired with 1h (1wk->1d->1h)
2023-08-13 10:54:57 +01:00
ValueRaider
e1f94ed337 Merge pull request #1654 from ranaroussi/fix/circular-import
Fix circular import in utils.py
2023-08-05 12:56:15 +01:00
Value Raider
93a7ee6161 Fix circular import in utils.py
Commit a4d7d6 introduced a circular import into utils.py
2023-08-05 12:54:12 +01:00
Value Raider
5b0cb60cf5 Bump version to 0.2.27 2023-08-03 21:24:07 +01:00
ValueRaider
1a97c22874 Merge pull request #1635 from ranaroussi/hotfix/prices-events-merge
Fix merging 1d-prices with out-of-range divs/splits
2023-08-03 21:20:29 +01:00
ValueRaider
b0de31da63 Merge pull request #1648 from ranaroussi/hotfix/tkr-tz-already-in-cache
Fix multithread error 'tz already in cache'
2023-08-03 18:02:51 +01:00
Value Raider
cc87608824 Fix multithread error 'tz already in cache' 2023-08-02 19:29:06 +01:00
ValueRaider
6c1e26093c Merge pull request #1633 from ranaroussi/feature/price-repair-improvements
Improve price repair
2023-08-01 20:03:29 +01:00
ValueRaider
e8fdd12cb1 Merge branch 'dev' into feature/price-repair-improvements 2023-07-31 18:30:00 +01:00
Value Raider
93b6e024da Improve bad-split-repair on multiday intervals
Improve bad-split-repair on multiday intervals
Switch some repair log msgs from warning -> info
2023-07-31 18:19:42 +01:00
ValueRaider
d5282967ce Merge pull request #1606 from DanielGoldfarb/main
option_chain() return underlying data that comes with the options data
2023-07-27 15:10:13 +01:00
ValueRaider
9908c1ff48 Merge pull request #1638 from ricardoprins/dev
PEP 8 changes + minor performance improvements
2023-07-26 17:49:25 +01:00
Ricardo Prins
a4d7d6c577 PEP 8 changes + minor performance improvements 2023-07-25 21:01:46 -06:00
ValueRaider
f9080c22a5 Merge pull request #1630 from ricardoprins/dev
Adjust PEP 8 + minor improvements + f-strings in base.py
2023-07-25 16:50:01 +01:00
Ricardo Prins
32e1d479b1 Remove last _pd 2023-07-23 18:37:09 -06:00
Ricardo Prins
5729ce3cb6 Remove .formats and adjust imports 2023-07-23 18:28:30 -06:00
ValueRaider
d0b2070036 Fix merging 1d-prices with out-of-range divs/splits 2023-07-23 15:20:57 +01:00
ValueRaider
688120cab7 Merge pull request #1632 from ranaroussi/fix/history-30m-typo
Fix typo in Ticker.history(30m)
2023-07-23 13:18:35 +01:00
ValueRaider
4a1e1c4447 Merge branch 'dev' into feature/price-repair-improvements 2023-07-22 20:34:56 +01:00
ValueRaider
f99677ed1e Fix typo in Ticker.history(30m) 2023-07-22 16:52:50 +01:00
ValueRaider
6a613eb114 Improve price repair
Several improvements to price repair

Repair 100x and split errors:
- Handle stocks that recently suspended - use latest ACTIVE trading as baseline
- Improve error identification:
  - Restrict repair to no older than 1 year before oldest split
  - To reduce false positives when checking for multiday split errors,
    only analyse 'Open' and 'Close' and use average change instead of nearest-to-1
  - For weekly intervals reduce threshold to 3x standard deviation (5x was too high),
    and for monthly increase to 6x
  - For multiday intervas, if errors only detected in 1 column then assume false positive => ignore

Repair missing div-adjust:
- Fix repair of multiday intervals containing dividend

Price reconstruction:
- Move to after repairing 100x and split errors, so calibration works properly
- Fix maximum depth and reduce to 1
- Restrict calibration to 'Open' and 'Close', because 'Low' and 'High' can differ significantly between e.g. 1d and day-of-1h

Miscellaneous:
- Deprecate repair='silent', the logging module handles this
- Improve tests for 100x and split errors
- New test for 'repair missing div adjust'
2023-07-22 15:09:38 +01:00
Ricardo Prins
0503240973 Adjust PEP 8 + minor improvements + f-strings 2023-07-21 20:33:47 -06:00
ValueRaider
ae6c05fa74 Merge pull request #1628 from ricardoprins/dev
Adjust PEP 8 + minor improvement
2023-07-21 22:35:21 +01:00
Ricardo Prins
aa9a0286a1 Add missing comment to test_dailyWithEvents 2023-07-21 14:25:46 -06:00
ValueRaider
ddf0cf19cd Bump version to 0.2.26 2023-07-21 12:56:10 +01:00
Ricardo Prins
a2bde88c36 Adjust PEP 8 + minor improvement 2023-07-20 22:44:36 -06:00
ValueRaider
1bd819ac4d Merge pull request #1371 from ranaroussi/hotfix/proxy
Fix proxy arg passthrough
2023-07-21 01:01:14 +01:00
ValueRaider
1b9fc5f12f Merge pull request #1625 from ricardoprins/main
Bump requests to 2.31 and removes cryptography.
2023-07-21 00:59:46 +01:00
Ricardo Prins
274f309052 Bump requests to 2.31 and removes cryptography. 2023-07-20 17:17:44 -06:00
ValueRaider
edac283a60 Merge pull request #1623 from ranaroussi/bug-report-yaml
Fix yaml issue rendering
2023-07-19 18:21:52 +01:00
ValueRaider
781fad501f Merge branch 'main' into bug-report-yaml 2023-07-19 18:21:44 +01:00
ValueRaider
39527d24d4 Fix yaml issue template rendering 2023-07-19 18:21:04 +01:00
ValueRaider
45f1c88460 yaml issue template - escape some backticks 2023-07-19 18:09:20 +01:00
ValueRaider
7d638e1040 Merge pull request #1613 from ranaroussi/bug-report-yaml
Convert issue template to yaml
2023-07-19 18:05:40 +01:00
ValueRaider
97b13dfa8c Convert issue template to yaml + improve 2023-07-19 18:01:47 +01:00
ValueRaider
693565a85b Bump version to 0.2.25 2023-07-18 13:45:55 +01:00
ValueRaider
957051e0e8 Merge pull request #1605 from ranaroussi/dev
sync dev -> main
2023-07-18 12:02:19 +01:00
ValueRaider
bd81ebb4e9 Merge pull request #1611 from ricardoprins/main
[BUG] Fix failure when using single ISIN as a ticker (#1525)
2023-07-18 10:55:18 +01:00
ValueRaider
46f53f9957 Port proxy fix to relocated 'FastInfo' 2023-07-17 18:34:00 +01:00
ValueRaider
056b84d8fe Merge branch 'main' into hotfix/proxy 2023-07-17 18:29:04 +01:00
Ricardo Prins
835dbd9629 Fix failure when using single ISIN as a ticker 2023-07-17 08:49:39 -06:00
ValueRaider
07a4594455 Dev version 0.2.25b1 2023-07-14 21:55:29 +01:00
Daniel Goldfarb
736c03ac5b options_chain() return underlying data that comes with the options data 2023-07-14 15:17:17 -04:00
ValueRaider
adfa2e9beb Merge pull request #1604 from ranaroussi/main
sync main -> dev
2023-07-14 20:11:45 +01:00
ValueRaider
b286797e8c Bump version to 0.2.24 2023-07-14 15:52:33 +01:00
ValueRaider
b306bef350 Merge pull request #1603 from ranaroussi/hotfix/info-missing-values
Fix info[] missing values
2023-07-14 15:51:23 +01:00
ValueRaider
61c89660df Optimise info fetch, improve test 2023-07-14 15:29:55 +01:00
Value Raider
31af2ab1d5 Fix recently-fixed info[] missing data 2023-07-13 22:20:42 +01:00
Value Raider
21c380fa61 Bump version to 0.2.23 2023-07-13 20:54:56 +01:00
ValueRaider
e0000cd787 Merge pull request #1595 from signifer-geo/bug20230714
Update quote.py
2023-07-13 20:51:33 +01:00
signifer-geo
11d43eb1a1 Update quote.py
dead code deleted
2023-07-14 04:29:59 +09:00
signifer-geo
509a109f29 Update quote.py
It fixes the error: unauthorized, invalid crumb
2023-07-14 03:11:24 +09:00
ValueRaider
b0639409a3 Merge pull request #1586 from ranaroussi/improve-readme
Emphasise API on Wiki
2023-07-10 15:26:31 +01:00
ValueRaider
ed10feee9a Merge pull request #1584 from lucas03/lukas/start-date-docs
update start parameter docstring
2023-07-06 21:20:17 +01:00
ValueRaider
aba81eedc2 Emphasise API on Wiki
More emphasis that user should review the Wiki for the full API for download() and Ticker.history()
2023-07-06 21:15:41 +01:00
Lukas Vojt
d424d027ac update docstrings for start parameter
requested here
https://github.com/ranaroussi/yfinance/pull/1576#issuecomment-1616599633
2023-07-06 08:17:07 +00:00
ValueRaider
9268fcfa76 Merge pull request #1545 from SnoozeFreddo/main
fix: Readme cache-ratelimit. Limiter parenthesis was never closed
2023-06-27 13:54:08 +01:00
ValueRaider
711e1138d3 Merge pull request #1576 from lucas03/lukas/start-date
fix start date on history
2023-06-27 12:32:17 +01:00
Lukas Vojt
0789b690a4 fix: start year on history
timestamp of 1900 is older than 100 years,
so yahoo responds with error:

GDEVW: 1d data not available for startTime=-2208994789 and
endTime=1687780922. Only 100 years worth of day granularity data are
allowed to be fetched per request.

this should fix it,
something similar was proposed here:
https://github.com/ranaroussi/yfinance/pull/648

 # Please enter the commit message for
your changes. Lines starting
2023-06-26 18:43:35 +02:00
ValueRaider
6055566de8 Bump version to 0.2.22 2023-06-24 19:42:18 +01:00
ValueRaider
398a19a855 Merge pull request #1574 from ranaroussi/hotfix/sql-db-error
Fix unhandled 'sqlite3.DatabaseError'
2023-06-24 19:40:51 +01:00
ValueRaider
e771cfabb6 Fix unhandled 'sqlite3.DatabaseError'
... also move '_TzCacheException' logging from level ERROR to INFO, because users don't need to know
2023-06-24 16:57:06 +01:00
ValueRaider
5b676f803b Bump version to 0.2.21 2023-06-21 23:15:37 +01:00
ValueRaider
eb5c50d5c7 Merge pull request #1569 from ranaroussi/dev
sync dev -> main
2023-06-21 23:13:49 +01:00
ValueRaider
1cb0b215c4 Merge branch 'main' into dev 2023-06-21 23:13:12 +01:00
ValueRaider
50dcb2ce5a Merge pull request #1568 from ranaroussi/fix/financials
Fix financials tables
2023-06-21 23:07:50 +01:00
ValueRaider
1ce9ce2784 Fix financials ; Remove broken decryption & scraping 2023-06-21 14:49:16 +01:00
ValueRaider
cd4816e289 Post-merge tidy price-repair logging 2023-06-20 17:46:06 +01:00
ValueRaider
27e9ce7542 Merge pull request #1543 from ranaroussi/feature/fix-Yahoo-bad-div-split-adjustments
Price repair update: fix Yahoo messing up dividend and split adjustments
2023-06-20 17:41:16 +01:00
ValueRaider
02c1c60f3b Merge branch 'dev' into feature/fix-Yahoo-bad-div-split-adjustments 2023-06-20 17:39:19 +01:00
ValueRaider
27ea9472c1 Price repair improvements
Improve split-repair of multi-day intervals. Because split error can occur within a multi-day interval, e.g. mid-way through week, need to repair each OHLC column separately

Increase robustness of repair 'Adj Close'

Limit price-repair recursion depth to 2
2023-06-20 17:37:19 +01:00
ValueRaider
801f58790a Merge pull request #1562 from ranaroussi/fix/logging-behaviour
Fix logging behaviour
2023-06-19 23:26:14 +01:00
ValueRaider
080834e3ce Update README#logging ; remove 'info fixed' message 2023-06-19 23:25:36 +01:00
ValueRaider
4e7b2094d0 Logging: improve appearance, fix propagation
Various important changes to yfinance logging:

Remove handler from YF logger as breaks standard logging practice.

Improve 'download' log message grouping

Move custom DEBUG log formatting into 'yf.enable_debug_mode()', which if called adds the handler. The custom DEBUG log formatting is:
- 'MultiLineFormatter' improves appearance of multi-line messages in any logging mode. Adds leading indent to all lines below first, to align with the first line which is indented by %levelname%.
- Whitespace padding is added to end of %levelname% string up to 8 characters. This resolves different level strings e.g. INFO & DEBUG creating different indents. But this only automatically applies to first line of message - for multi-line messages, 'MultiLineFormatter' extracts amount of padding and applies to all other lines.
- Add leading indent proportional to function depth, because DEBUG generates a lot of messages particularly when repairing price data - same function can be called recursively. Implemented in 'IndentLoggerAdapter', a simple wrapper around the logger object.
- 'log_indent_decorator' inserts 'Entering/Exiting %function%' debug messages, helps organise.
2023-06-19 19:42:02 +01:00
ValueRaider
c72e04bf55 Merge pull request #1567 from ranaroussi/fix/future-event-merge
Fix merge future div/split into prices
2023-06-19 11:56:48 +01:00
ValueRaider
abbe4c3a2f Fix merge future div/split into prices 2023-06-19 11:41:53 +01:00
ValueRaider
9e21b85043 Merge branch 'dev' into feature/fix-Yahoo-bad-div-split-adjustments 2023-06-17 12:49:01 +01:00
ValueRaider
b44917b7f9 Improve price-repair of 1d 'Adj Close'
Main change is fixing price-repair of 1d 'Adj Close'. 1d repair uses 1h data, but 1h is never div-adjusted. For correct 1d 'Adj Close', extract div-adjustment from the good 1d data, and calculate it for any bad 1d data. A new unit test ensures correctness.

Other changes:
- bug fix in split-repair logic to handle price=0
- improve unit test coverage on price dividend
- add 1wk interval to split-repair unit test
2023-06-16 14:06:25 +01:00
ValueRaider
6f78dd6e6b Fix dumb bug 2023-06-12 17:05:37 +01:00
ValueRaider
593dc8fcee Use new split-repair logic to also fix 100x error
Stumbled upon another type of 100x price error - Yahoo may switch a symbol from e.g. cents -> $ on some recent date, so recent prices are 100x different. The new split-repair is perfect for this - set change to 100 and ignore 'Volume'.
2023-06-12 14:46:58 +01:00
ValueRaider
b94baa4cc5 Merge pull request #1552 from adityanparikh/fix/timedelta
Fix timedelta bug
2023-06-08 21:26:31 +01:00
ValueRaider
1a054135fb Merge pull request #1553 from ranaroussi/fix/events-merging
Fix prices-events merge
2023-06-08 21:18:14 +01:00
ValueRaider
4e2253a406 Fix prices-events merge
An out-of-range dividend was breaking merge with 1mo-prices, so fixed that. Also replaced the mega-loop with Numpy, much clearer now. Improved its tests.
2023-06-08 21:16:05 +01:00
Aditya Parikh
9af7ec0a4e Fix timedelta bug
Fix for exception using _pd.Timedelta function with unit='d' parameter for  Pandas v1.4.4
2023-06-08 16:15:03 -04:00
Value Raider
8624216e21 Bump version to 0.2.20 2023-06-07 16:51:17 +01:00
ValueRaider
954e71d19c Update action versions in python-publish.yml
Recent release action generated deprecated error: "Node.js 12 actions are deprecated. Please update the following actions to use Node.js 16: actions/checkout@v2, actions/setup-python@v2."

So simply increasing versions to match latest GitHub usage docs, hopefully works.
2023-06-07 16:48:11 +01:00
ValueRaider
5124059422 Bump version to 0.2.19 2023-06-07 13:28:32 +01:00
ValueRaider
d18cd6f42f Merge pull request #1549 from ranaroussi/dev
dev -> main
2023-06-07 13:23:39 +01:00
ValueRaider
c20211a06c Merge pull request #1547 from bveber/dev 2023-06-06 23:05:40 +01:00
bveber
cdfe7d0d2d add session to download 2023-06-06 01:06:18 -05:00
ValueRaider
e57647c1d7 Stock split repair: bug fixes & more testing 2023-06-03 20:49:35 +01:00
Konstantinos Ftikas
762abd8bba fix: Readme cache-ratelimit. Limiter parenthesis was never closed
The example in the docs will not work out of the box due to a syntax error.
2023-06-03 14:19:31 +02:00
ValueRaider
d1ea402792 Price repair improvement: fix stupid bug 2023-06-01 15:56:54 +01:00
ValueRaider
65f65b1776 Price repair improvement: fix price Low=High errors on stock split day
Yahoo often messes up price data on stock split day - all equal or simply missing.
Impossible that price didn't move/trade on stock split day.
2023-05-31 21:51:04 +01:00
ValueRaider
9388c29207 Price repair improvement: fix stock split adjustment missing from pre-split data 2023-05-31 21:41:05 +01:00
ValueRaider
9f91f4b180 Price repair improvement: fix missing dividend adjustment 2023-05-31 21:40:16 +01:00
ValueRaider
cac616a24c Dev version 0.2.19b4 2023-05-25 11:09:31 +01:00
ValueRaider
72a9e45e56 Merge pull request #1541 from ranaroussi/fix/download-logging
Bugfix in `download` logging tracebacks & boost tests
2023-05-25 10:58:52 +01:00
ValueRaider
4802199ae7 Bugfix in download logging tracebacks & boost tests
New logging in `download` stores the tracebacks, but the logic was faulty, this fixes that.
Also improves error handling in `download`.
Unit tests should have detected this so improved them:
- add/improve `download` tests
- disable tests that require Yahoo decryption (because is broken)
- fix logging-related errors
- improve session use
2023-05-24 13:19:39 +01:00
ValueRaider
d9bfd29113 Delete 'Feature request' issue template - can't have nice things 2023-05-23 17:11:31 +01:00
ValueRaider
4711aab7b3 Merge pull request #1536 from ranaroussi/hotfix/tz-cache-migrate-error-again
Fix corrupt tkr-tz-csv halting code (again)
2023-05-23 16:44:35 +01:00
ValueRaider
30d20c1206 Fix corrupt tkr-tz-csv halting code (again) 2023-05-23 16:34:50 +01:00
ValueRaider
5c565c8934 bug_report.md: add instruction to post debug log
bug_report.md: add instruction to post debug log. Plus some minor edits.
2023-05-17 18:44:52 +01:00
ValueRaider
2fff97290b Merge pull request #1528 from ranaroussi/fix/tz-cache-migrate-error
Fix corrupt tkr-tz-csv halting code
2023-05-17 16:59:15 +01:00
ValueRaider
62ca5ab6be Fix corrupt tkr-tz-csv halting code 2023-05-17 15:05:38 +01:00
ValueRaider
83b177b7fb README.md - note on installing betas 2023-05-12 12:11:14 +01:00
ValueRaider
e8b99cb4e6 Dev version 0.2.19b3 2023-05-11 14:04:04 +01:00
ValueRaider
503d234020 Dev version 0.2.19b2 - add missing file 2023-05-11 14:03:38 +01:00
ValueRaider
144efd3b08 Dev version 0.2.19b2 2023-05-11 13:52:41 +01:00
ValueRaider
80fc91ffa9 Merge pull request #1523 from ranaroussi/fix/price-fixes
Price fixes
2023-05-11 13:51:03 +01:00
ValueRaider
9821197fd1 Merge pull request #1522 from ranaroussi/fix/logging-messages
Improve logging messages
2023-05-11 13:50:45 +01:00
ValueRaider
45b5cac33b Improve logging messages
Improve logging messages related to price data fetches:
- fix 'debug is deprecated' msg
- append user args to 'may be delisted' msg - interval & dates/period
- improve formatting of 'cannot reconstruct' msg
- hide errors in 'history()' while accessing 'fast_info[]'
2023-05-10 14:47:58 +01:00
ValueRaider
d755b8c7ff Fix 'history()' edge cases
Fix merging prices & events if prices empty.
If user requested price repair, ensure 'Repaired?' column always present.
2023-05-10 14:44:50 +01:00
ValueRaider
ab1042b4c9 Dev version 0.2.19b1 2023-05-04 22:14:34 +01:00
ValueRaider
8172fc02d2 Merge pull request #1514 from ranaroussi/feature/optimise-history
Optimise Ticker.history() - up to 2x faster
2023-05-04 22:08:40 +01:00
ValueRaider
836082280b Merge branch 'dev' into feature/optimise-history 2023-05-04 22:08:28 +01:00
ValueRaider
6a98c2eda6 Merge pull request #1493 from ranaroussi/feature/error-reporting
Deprecate 'debug' arg, improve 'logging' use
2023-05-04 22:06:54 +01:00
ValueRaider
46f55c8983 Add debug logging to 'history()' ; Improve logger fmt 2023-05-04 22:04:39 +01:00
ValueRaider
b025fef22c Optimise Ticker.history() - up to 2x faster
format_history_metadata() is expensive. Improvements:
- only perform full formatting if user requests metadata
- when pruning prepost data, only format 'tradingPeriods' entry of metadata

Other small optimisations to several internal prices processing methods.

Speedups:
dat.history(period='1wk', interval='1h', prepost=True)  # 2x
dat.history(period='1mo', interval='1h', prepost=True)  # 1.46x
dat.history(period='1wk', interval='1h')  # 1.15x
dat.history(period='1mo', interval='1h')  # 1.13x
dat.history(period='1y', interval='1d')  # 1.36x
dat.history(period='5y', interval='1d')  # 1.13x
2023-04-30 00:35:08 +01:00
ValueRaider
b96319dd64 Merge pull request #1504 from ranaroussi/hotfix/sql-exception
Fix timezone cache error: IntegrityError('NOT NULL constraint failed: kv.key')
2023-04-26 21:29:33 +01:00
ValueRaider
74b88dc62c Fix IntegrityError in timezone cache 2023-04-26 21:27:31 +01:00
ValueRaider
e3778465d8 Merge branch 'dev' into feature/error-reporting 2023-04-22 16:02:56 +01:00
ValueRaider
f82177ea2e Improve download() logging - group errors & tracebacks for cleaner STDOUT 2023-04-16 21:57:04 +01:00
ValueRaider
d30a2a0915 README.md: update 'News' 2023-04-16 21:29:57 +01:00
ValueRaider
142b1f3eb4 Merge pull request #1499 from ranaroussi/main
sync main -> dev
2023-04-16 19:08:50 +01:00
ValueRaider
afad7fcf0b Bump version to 0.2.18 2023-04-16 19:03:08 +01:00
ValueRaider
0baedbe4f5 Merge pull request #1498 from ranaroussi/hotfix/tz-cache-migrate-error
Fix handling Pandas parsing error during TZ-csv-cache migrate
2023-04-16 19:00:50 +01:00
ValueRaider
2c3c3dc8a9 Merge pull request #1496 from ranaroussi/hotfix/fast-info-np-not-found
Fix '_np not found', tweak 'info[] fixed' message
2023-04-16 18:59:38 +01:00
ValueRaider
8585dda77a Fix handling Pandas parsing error during TZ-csv-cache migrate 2023-04-16 15:09:28 +01:00
ValueRaider
3eb60fbd4a Fix '_np not found', tweak 'info[] fixed' message 2023-04-16 10:37:25 +01:00
ValueRaider
d3e2e71a6e Improve logging behaviour, particulary download()
- Use same logger across all files
- download():
  - write tracebacks to DEBUG
  - deprecate 'show_errors' argument
2023-04-15 17:29:07 +01:00
ValueRaider
4937c933a2 Deprecate 'debug' arg, improve 'logging' use 2023-04-15 16:47:39 +01:00
ValueRaider
045cd45893 Bump version to 0.2.17 2023-04-10 21:55:21 +01:00
ValueRaider
6d52cb6e3a Merge pull request #1488 from steven9909/fix_localize
Fix tzinfo missing attribute
2023-04-10 21:51:54 +01:00
steven9909
a24c0e1391 fix tzinfo missing attribute
tzinfo does not have a localize attribute so it is replaced with timestamp in UTC
2023-04-10 16:04:58 -04:00
ValueRaider
1e941fc86a Merge branch 'main' into dev 2023-04-09 23:45:37 +01:00
ValueRaider
0b52e8f118 Bump version to 0.2.16 2023-04-09 23:42:50 +01:00
ValueRaider
d45bed3d53 Fix 'fast_info deprecated' msg appearing at Ticker() init 2023-04-09 23:41:44 +01:00
ValueRaider
4152f7c897 Bump version to 0.2.15 2023-04-09 21:07:16 +01:00
ValueRaider
e7a3848f69 Merge pull request #1477 from ranaroussi/feature/price-repair-tweaks
Price repair: add 'Repaired?' column, and a bugfix
2023-04-09 21:01:49 +01:00
ValueRaider
fc4350e463 Merge pull request #1480 from kennykos/get_full_info
'info' fetch now gets same data as scrape
2023-04-09 21:01:34 +01:00
ValueRaider
13556afd90 README.md: reorganise & link to 'How to contribute' 2023-04-07 12:21:01 +01:00
ValueRaider
3d29ced428 Merge pull request #1474 from garrettladley/leverage-dict-and-list-comps
Leverage dict & list comprehensions in yfinance/tickers.py
2023-04-06 13:26:08 +01:00
Value Raider
6a63ce9e15 Demote 'fast_info'
Demote 'fast_info':
- inform user can revert to 'info'
- remove from README
- relocate class from base.py -> quote.py
2023-04-06 12:21:57 +01:00
garrettladley
2fe5a0a361 leveraged dict & list comps in yfinance/tickers.py 2023-04-05 18:55:47 -04:00
kennykos
63699a6aad 'info' fetch now gets same data as scrape
* Changed base url to "https://query2.finance.yahoo.com/v10/finance/quoteSummary"
* instead of just getting the quote, we now get
	* ```
	   items = ['summaryProfile', 'financialData', 'quoteType',
                    'defaultKeyStatistics', 'assetProfile', 'summaryDetail']
          ```
	which is the same as in the scrape function
2023-04-05 18:23:36 +01:00
Value Raider
a649b40dc9 Price repair: add 'Repaired?' column, and a bugfix
Price repair changes:
- if user requests price repair, add 'Repaired?' bool column showing what rows were repaired.
- fix price repair requesting <1d data beyond Yahoo's limit.
- fix logger messages
2023-04-03 21:27:04 +01:00
ValueRaider
a01edee4fa Merge pull request #1476 from ranaroussi/main
main -> dev
2023-04-03 21:20:50 +01:00
Value Raider
5367f62bd7 Bump version to 0.2.14 2023-03-25 11:39:21 +00:00
ValueRaider
27cb90c596 Merge pull request #1461 from qianyun210603/main
Add failback for decryption error in info interface
2023-03-25 11:33:27 +00:00
BookSword
6c2682654a Fetch 'info' dict via API 2023-03-24 18:04:07 +00:00
Value Raider
e89e190d11 Merge branch 'main' into dev 2023-03-21 19:05:56 +00:00
ValueRaider
a236270389 Merge pull request #1457 from ranaroussi/fix/price-fixes-various
Various fixes to price data processing
2023-03-21 18:59:13 +00:00
Value Raider
ef1205388c Bump version to 0.2.13 2023-03-21 18:56:32 +00:00
Value Raider
bb477989d4 Fix price-events merge when occurred pre-market 2023-03-21 18:52:35 +00:00
ValueRaider
478dc0a350 Merge pull request #1452 from ranaroussi/hotfix/prices-merge-events
Fix filtering events older than prices for merging
2023-03-21 18:16:29 +00:00
Value Raider
b5dca4941a Order history_metadata['tradingPeriods'] DF sensibly 2023-03-20 21:18:53 +00:00
Value Raider
6b71ba977c Various fixes to price data processing
- move drop-duplicates to before repair
- fix 'format_history_metadata()' processing 'regular' column
- fix Pandas & Numpy warnings
2023-03-20 21:10:45 +00:00
ValueRaider
195a7aa304 Merge pull request #1455 from mppics/fix/aggregate_capital_gains
Adding fix and test for aggregating Capital Gains
2023-03-18 17:24:53 +00:00
Matt Piccoli
a58d7456fe Adding fix and test for aggregating Capital Gains 2023-03-18 12:57:26 -04:00
ValueRaider
1edeaf07dc Merge pull request #1448 from ivan23kor/feature/clarify-end-argument
Clarify that interval is [start; end) in docstrings
2023-03-09 22:04:58 +00:00
Ivan Korostelev
7f04a9dcb6 Clarify that interval is [start; end) in docstrings 2023-03-09 14:27:21 -07:00
ValueRaider
7b95f554bd README: fix rate-limiting example 2023-02-21 12:24:35 +00:00
ValueRaider
6c70b866c7 Merge pull request #1423 from flaviovs/no-print
No print
2023-02-20 20:07:23 +00:00
Value Raider
bd696fb4db Beta version 0.2.13b1 2023-02-17 17:04:39 +00:00
Value Raider
d13aafa633 Replace more prints with logging, mostly in 'price repair' 2023-02-17 12:01:11 +00:00
Flávio Veloso Soares
00823f6fa6 Remove redundant logging text 2023-02-16 16:53:33 -08:00
Flávio Veloso Soares
21fdba9021 Replace warnings print() with warnings.warn(...) calls 2023-02-16 16:53:33 -08:00
Flávio Veloso Soares
972547ca8c Replace prints with logging module 2023-02-16 16:53:33 -08:00
ValueRaider
23b400f0fb Merge pull request #1421 from ranaroussi/fix/missing-price-history-errors
Improve handling missing price history
2023-02-16 14:22:10 +00:00
Value Raider
ca8c1c8cb4 Bump version to 0.2.12 2023-02-16 12:01:25 +00:00
ValueRaider
6b8b0d5c86 Merge pull request #1422 from ranaroussi/hotfix/disable-decrypt-fail-msg
Disable annoying 'backup decrypt' msg
2023-02-16 12:00:16 +00:00
Value Raider
952a04338f Disable annoying 'backup decrypt' msg 2023-02-15 16:46:55 +00:00
Value Raider
a1a385196b Improve handling missing price history
Fix fast_info[] dying if metadata incomplete/missing ; Price repair fix when no fine data available ; Fix _fix_unit_mixups() report
2023-02-14 17:31:14 +00:00
ValueRaider
62a442bd15 Update yahoo-keys.txt 2023-02-14 00:06:06 +00:00
ValueRaider
a0046439d1 Merge pull request #1400 from ranaroussi/feature/improve-performance
Optimise recent new features in `history`
2023-02-12 14:58:36 +00:00
ValueRaider
63a8476575 Merge pull request #1417 from ranaroussi/main
main -> dev
2023-02-12 14:56:19 +00:00
ValueRaider
e96f4f3cc0 Update yahoo-keys.txt 2023-02-12 09:57:25 +00:00
ValueRaider
cd5d0dfc3b Bump version to 0.2.11 2023-02-10 16:59:20 +00:00
ValueRaider
ece41cdb06 Merge pull request #1411 from sdeibel/main
Fix format_history_metadata for some symbols
2023-02-10 16:30:03 +00:00
ValueRaider
c362d54b1a Fix other metadata accesses + tests 2023-02-09 19:41:50 +00:00
Stephan Deibel
543e4fe582 Fix format_history_metadata for some symbols
Fix format_history_metadata when firstTradeDate is None, as is the case for QCSTIX and probably others.
2023-02-09 13:46:52 -05:00
ValueRaider
f8aab533ba Merge branch 'main' into hotfix/proxy 2023-02-08 13:52:34 +00:00
ValueRaider
53fca7016e Bump version to 0.2.10 2023-02-07 22:05:17 +00:00
ValueRaider
4b6529c3a5 Merge pull request #1406 from ranaroussi/dev
dev -> main
2023-02-07 22:03:20 +00:00
ValueRaider
8957147926 Merge branch 'main' into dev 2023-02-07 22:02:46 +00:00
ValueRaider
4c7392ed17 Merge pull request #1403 from ranaroussi/fix/decrypt-keys
Fix decrypt keys
2023-02-07 21:55:33 +00:00
ValueRaider
0efda4f5af Fix filtering events older than prices for merging 2023-02-07 21:45:35 +00:00
ValueRaider
508de4aefb Dev version 0.2.10b3 2023-02-07 14:09:08 +00:00
ValueRaider
3d39992280 Add resilience to price repair
When calibrating price repair, use weighted average to estimate stock split ratio, is more resilient
2023-02-07 14:07:08 +00:00
ValueRaider
b462836540 Merge pull request #1385 from ranaroussi/fix/download-tz-behaviour
Restore original download() timezone handling
2023-02-07 13:16:03 +00:00
ValueRaider
2795660c28 Add a 5th backup key 2023-02-07 13:10:03 +00:00
ValueRaider
3dc87753ea Fix _get_decryption_keys_from_yahoo_js() returning '' 2023-02-07 13:09:49 +00:00
ValueRaider
645cc19037 Merge pull request #1379 from ranaroussi/feature/improve-decrypt
Add another backup decrypt option
2023-02-06 22:24:22 +00:00
ValueRaider
86d6acccf7 Fix dumb bugs in price repair - 1 more 2023-02-05 18:17:47 +00:00
ValueRaider
0f5db35b6e Optimise Ticker._reconstruct_intervals_batch() (slightly) 2023-02-05 18:16:08 +00:00
ValueRaider
7c6742a60a Optimise Ticker._fix_unit_mixups() 2023-02-05 15:15:56 +00:00
ValueRaider
4fa32a98ed Merge pull request #1397 from Matt-Seath/dev
Catch TypeError Exception
2023-02-05 13:49:48 +00:00
ValueRaider
36ace8017d Optimise Ticker._fix_zeroes() 2023-02-05 13:46:57 +00:00
ValueRaider
5cdc78f479 Merge pull request #1398 from vidalmarco/patch-1
get_shares_full does not work with proxy
2023-02-05 10:59:33 +00:00
Marco Vidal
ba634fad0e get_shares_full does not work with proxy
Error: "Yahoo web request for share count failed" 
updated cache_get call by adding proxy parameter and by calling it by keyword
2023-02-05 09:17:22 +01:00
Matt Seath
35f4071c0b Catch TypeError Exception
Addresses recent issue where calling Ticker.info would occasionally result in a TypeError Exception at line 287.
2023-02-05 11:49:40 +10:00
ValueRaider
8a5ca71f52 Fix holders.py proxy pass-through 2023-02-05 00:06:49 +00:00
ValueRaider
ead0bce96e Optimise format_history_metadata() 2023-02-04 22:56:49 +00:00
ValueRaider
86b00091a9 Fix dumb bugs in price repair 2023-02-02 21:57:55 +00:00
ValueRaider
141ce7e471 Fix proxy + cache_get. Improve error propagation 2023-02-01 21:19:54 +00:00
ValueRaider
4eae728a06 Potential fix for proxy - enable #2 2023-02-01 19:17:18 +00:00
ValueRaider
2d6b6b26ed Potential fix for proxy - enable 2023-02-01 19:04:47 +00:00
ValueRaider
ec3dfaf305 Potential fix for proxy - revert 2023-02-01 18:10:45 +00:00
ValueRaider
e89d390824 Potential fix for proxy 2023-02-01 18:09:51 +00:00
ValueRaider
563a1a3448 Add Ticker test for proxy 2023-02-01 17:28:57 +00:00
ValueRaider
2e6d3d0e60 Fix proxy in 'history()' 2023-02-01 17:06:23 +00:00
ValueRaider
2a2928b4a0 Fix 'tradingPeriods' parsing when empty - 0.2.10b2 2023-02-01 13:31:54 +00:00
ValueRaider
d47133e5bf Dev version 0.2.10b1 2023-01-31 22:12:11 +00:00
ValueRaider
8f0c58dafa Dev version 0.2.10b0 2023-01-31 22:02:41 +00:00
ValueRaider
27a721c7dd Merge pull request #1380 from ranaroussi/fix/old-sqlite-error
Allow using sqlite3 < 3.8.2
2023-01-31 19:52:22 +00:00
ValueRaider
3e964d5319 Merge pull request #1383 from ranaroussi/fix/fast-info-prepost
Fix fast_info["previousClose"]
2023-01-31 19:51:46 +00:00
ValueRaider
84a31ae0b4 Merge pull request #1311 from ranaroussi/feature/prices-metadata-prune-prepost
Drop intraday intervals if in post-market but prepost=False
2023-01-31 19:50:00 +00:00
ValueRaider
891b533ec2 Drop intraday intervals if in prepost but prepost=False 2023-01-31 19:48:47 +00:00
ValueRaider
b9fb3e4979 Restore original download() tz handling: day/week/etc = ignore 2023-01-31 00:00:45 +00:00
ValueRaider
09342982a4 Add 'quoteType'. Improve handling tickers without trading 2023-01-30 23:53:06 +00:00
ValueRaider
da8c49011e fast_info: Fix previousClose & yearChange 2023-01-30 16:06:55 +00:00
ValueRaider
b805f0a010 Add another backup decrypt option 2023-01-29 23:09:45 +00:00
ValueRaider
5b0feb3d20 Fix tests 2023-01-29 16:53:26 +00:00
ValueRaider
ecbfc2957d bug_report: tighten language (again) 2023-01-29 13:58:02 +00:00
ValueRaider
e96248dec7 README: fix narrative ordering 2023-01-29 13:52:13 +00:00
ValueRaider
7d0045f03c README: simplify API overview with link to Wiki 2023-01-29 13:49:01 +00:00
ValueRaider
c3d7449844 Merge pull request #1289 from ranaroussi/fix/price-repair
Fix & improve price repair
2023-01-29 13:02:48 +00:00
ValueRaider
a4f11b0243 Fix price repair tests, remove unrelated changes 2023-01-29 13:01:54 +00:00
ValueRaider
1702fd0797 bug_report: tighten language 2023-01-29 00:54:27 +00:00
ValueRaider
464b3333d7 Allow using sqlite3 < 3.8.2 2023-01-29 00:34:46 +00:00
ValueRaider
685f2ec351 Merge branch 'dev' into fix/price-repair 2023-01-28 23:26:56 +00:00
ValueRaider
aad46baf28 price repair: Fix 'min_dt', add 'silent' mode 2023-01-28 23:14:28 +00:00
ValueRaider
a97db0aac6 README: add how-to for requests rate-limiting 2023-01-28 23:10:38 +00:00
ValueRaider
553bc5965a Fix proxy arg passthrough 2023-01-28 23:07:19 +00:00
ValueRaider
af5f96f97e Merge pull request #1368 from ranaroussi/fix/fast-info-camel-case
`fast_info` usability improvements
2023-01-28 22:28:42 +00:00
ValueRaider
a4bdaea888 fast_info: add camelCase, items() & values() 2023-01-28 22:27:51 +00:00
ValueRaider
ac5a9d2793 Merge pull request #1367 from ranaroussi/main
main -> dev
2023-01-27 22:09:59 +00:00
ValueRaider
b17ad32a47 Merge pull request #1366 from ranaroussi/doc/readme-explain-instability
README: comment on instability, tidy Ticker 'Quick start'
2023-01-27 18:31:32 +00:00
ValueRaider
af39855e28 README: comment on instability, tidy Ticker 'Quick start' 2023-01-27 17:36:25 +00:00
ValueRaider
ac6e047f0d Bump version to 0.2.9 2023-01-26 22:21:46 +00:00
ValueRaider
1e24337f29 Bump version to 0.2.8 2023-01-26 22:20:11 +00:00
ValueRaider
2cc82ae12f Merge pull request #1362 from ranaroussi/hotfix/fast-info-bugs
Ticker.fast_info: fix teething bugs
2023-01-26 22:03:06 +00:00
ValueRaider
d11f385049 Make fast_info JSON-serializable via toJSON() 2023-01-26 21:45:53 +00:00
ValueRaider
7377611e1f Add 'get(key, default)' to fast_info 2023-01-26 21:23:31 +00:00
ValueRaider
f3b5fb85c9 Remove exception raise from 'get_shares_full()' 2023-01-26 21:14:48 +00:00
ValueRaider
a4faef83ac 'fast_info' fixes: unusual symbols ; improve migration message ; 'regular_market_previous_close' 2023-01-26 21:02:18 +00:00
ValueRaider
e1184f745b Update yahoo-keys.txt 2023-01-26 17:06:03 +00:00
ValueRaider
fe630008e9 Bump version to 0.2.7 2023-01-26 17:03:00 +00:00
ValueRaider
b43072cf0a Merge pull request #1354 from ranaroussi/hotfix/rename-basic-info
Rename 'basic_info' -> 'fast_info'
2023-01-26 17:00:54 +00:00
ValueRaider
ad3f4cabc9 Improve 'get_shares_full()' error handling 2023-01-26 16:58:26 +00:00
ValueRaider
f70567872c Merge pull request #1353 from ranaroussi/hotfix/smart-decryption
Add decrypt key extraction from JS + GitHub backup
2023-01-26 16:44:23 +00:00
ValueRaider
a8ade72113 Rename 'basic_info' -> 'fast_info' ; Fix info tests 2023-01-26 16:36:25 +00:00
ValueRaider
1dcc8c9c8b Remove dead debug code 2023-01-26 14:57:15 +00:00
ValueRaider
dd5462b307 Add decrypt key extraction from JS + GitHub backup 2023-01-26 14:52:18 +00:00
ValueRaider
e39c03e8e3 Hardcode decrypt keys in GitHub for fix w/o PIP
`yfinance` will query this file via web request as a last resort. Avoids having to release a new PIP version just for a key update.
2023-01-26 14:20:03 +00:00
ValueRaider
9297504b84 Merge pull request #1346 from ranaroussi/main
main -> dev sync
2023-01-25 22:16:22 +00:00
ValueRaider
39c1ecc7a2 Improve price repair - reduce spam, improve data reliability
Extend 'reconstruct groups' to reduce Yahoo spam ; Extend fetch range to avoid first/last day irregularities ; Improve handling of 'max fetch days' Yahoo limit
2023-01-25 14:37:43 +00:00
ValueRaider
eb6d830e2a Fix repair volume=0 ; Tidy code 2023-01-21 23:00:30 +00:00
ValueRaider
2b0ae5a6c1 Remove 'repair_intervals' 2023-01-21 16:58:45 +00:00
ValueRaider
1636839b67 Handle request to reconstruct 1m 2023-01-20 00:13:28 +00:00
ValueRaider
65b97d024b Improve reporting 2023-01-20 00:13:02 +00:00
ValueRaider
197d2968e3 Add 'repair_intervals', rename 'repair'->'repair_prices' 2023-01-19 22:19:16 +00:00
ValueRaider
7460dbea17 If reconstructing 1d interval with 1h, always request prepost 2023-01-19 22:18:46 +00:00
ValueRaider
b49fd797fc Fix & improve price repair
Fix repair calibration & volume=0 repair ; Extend repair to sub-hour ; Avoid attempting repair of mostly-NaN days
2023-01-19 22:18:46 +00:00
ValueRaider
0ba810fda5 Improve 'history_metadata' formatting 2023-01-16 18:30:28 +00:00
62 changed files with 6502 additions and 3055 deletions

View File

@@ -1,42 +0,0 @@
---
name: Bug report
about: Create a report to help us improve
title: ''
labels: ''
assignees: ''
---
# READ BEFORE POSTING
### Are you up-to-date?
Upgrade to the latest version and confirm the issue/bug is still there.
`$ pip install yfinance --upgrade --no-cache-dir`
Confirm by running:
`import yfinance as yf ; print(yf.__version__)`
and comparing against [PIP](https://pypi.org/project/yfinance/#history).
### Does Yahoo actually have the data?
Are spelling ticker *exactly* same as Yahoo?
Visit `finance.yahoo.com` and confim they have your data. Maybe your ticker was delisted.
### Are you spamming Yahoo?
Yahoo Finance free service has limit on query rate dependent on request - roughly 500/minute for prices, 10/minute for info. Them delaying or blocking your spam is not a bug.
### Still think it's a bug?
Delete this default message and submit your bug report here, providing the following as best you can:
- Simple code that reproduces your problem
- Error message, with traceback if shown
- Info about your system:
- yfinance version
- operating system

89
.github/ISSUE_TEMPLATE/bug_report.yaml vendored Normal file
View File

@@ -0,0 +1,89 @@
name: Bug report
description: Report a bug in our project
labels: ["bug"]
body:
- type: markdown
attributes:
value: |
# !!! IMPORTANT !!! FOLLOW THESE INSTRUCTIONS CAREFULLY !!!
### Are you up-to-date?
Upgrade to the latest version: `$ pip install yfinance --upgrade --no-cache-dir`
Confirm latest version by running: `import yfinance as yf ; print(yf.__version__)` and comparing against [PyPI](https://pypi.org/project/yfinance/#history).
### Does Yahoo actually have the data?
Are you spelling symbol *exactly* same as Yahoo?
Then visit `finance.yahoo.com` and confirm they have the data you want. Maybe your symbol was delisted, or your expectations of `yfinance` are wrong.
### Are you spamming Yahoo?
Yahoo Finance free service has rate-limiting https://github.com/ranaroussi/yfinance/discussions/1513. Once limit hit, Yahoo can delay, block, or return bad data -> not a `yfinance` bug.
### Does issue already exist?
Use the search tool. Don't duplicate existing issues.
- type: markdown
attributes:
value: |
---
## Still think it's a bug?
Provide the following as best you can:
- type: textarea
id: summary
attributes:
label: "Describe bug"
validations:
required: true
- type: textarea
id: code
attributes:
label: "Simple code that reproduces your problem"
description: "Provide a snippet of code that we can copy-paste-run. Wrap code in Python Markdown code blocks for proper formatting (```` ```python ... ``` ````)."
validations:
required: true
- type: textarea
id: debug-log
attributes:
label: "Debug log"
description: "Run code with debug logging enabled and post the full output. IMPORTANT INSTRUCTIONS: https://github.com/ranaroussi/yfinance/tree/main#logging"
validations:
required: true
- type: textarea
id: bad-data-proof
attributes:
label: "Bad data proof"
description: "If you think `yfinance` returning bad data, provide your proof here."
validations:
required: false
- type: input
id: version-yfinance
attributes:
label: "`yfinance` version"
validations:
required: true
- type: input
id: version-python
attributes:
label: "Python version"
validations:
required: false
- type: input
id: os
attributes:
label: "Operating system"
validations:
required: false

View File

@@ -1,14 +0,0 @@
---
name: Feature request
about: Request a new feature
title: ''
labels: ''
assignees: ''
---
**Describe the problem**
**Describe the solution**
**Additional context**

View File

@@ -8,11 +8,11 @@ jobs:
deploy:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: actions/setup-python@v2
- uses: actions/checkout@v3
- uses: actions/setup-python@v4
with:
python-version: 3.x
- run: pip install -r requirements.txt
- run: pip install mkdocstrings==0.14.0
- run: pip install mkdocs-material
- run: mkdocs gh-deploy --force
- run: mkdocs gh-deploy --force

View File

@@ -13,9 +13,9 @@ jobs:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: actions/checkout@v3
- name: Set up Python
uses: actions/setup-python@v2
uses: actions/setup-python@v4
with:
python-version: '3.x'
- name: Install dependencies

13
.github/workflows/ruff.yml vendored Normal file
View File

@@ -0,0 +1,13 @@
name: Ruff
on:
pull_request:
branches:
- master
- main
- dev
jobs:
ruff:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
- uses: chartboost/ruff-action@v1

1
.gitignore vendored
View File

@@ -4,6 +4,7 @@ dist
yfinance.egg-info
*.pyc
.coverage
.idea/
.vscode/
build/
*.html

View File

@@ -1,6 +1,191 @@
Change Log
===========
0.2.38
------
Fix holders & insiders #1908
0.2.37
------
Small fixes:
- Fix Pandas warnings #1838 #1844
- Fix price repair bug, typos, refactor #1866 #1865 #1849
- Stop disabling logging #1841
0.2.36
------
Small fixes:
- Update README.md for better copy-ability #1823
- Name download() column levels #1795
- Fix history(keepna=False) when repair=True #1824
- Replace empty list with empty pd.Series #1724
- Handle peewee with old sqlite #1827
- Fix JSON error handling #1830 #1833
0.2.35
------
Internal fixes for 0.2.34
0.2.34
------
Features:
- Add Recommendations Trend Summary #1754
- Add Recommendation upgrades & downgrades #1773
- Add Insider Roster & Transactions #1772
- Moved download() progress bar to STDERR #1776
- PIP optional dependencies #1771
- Set sensible min versions for optional 'nospam' reqs #1807
Fixes
- Fix download() DatetimeIndex on invalid symbols #1779
- Fix invalid date entering cache DB #1796
- Fix Ticker.calendar fetch #1790
- Fixed adding complementary to info #1774
- Ticker.earnings_dates: fix warning "Value 'NaN' has dtype incompatible with float64" #1810
- Minor fixes for price repair and related tests #1768
- Fix price repair div adjust #1798
- Fix 'raise_errors' argument ignored in Ticker.history() #1806
Maintenance
- Fix regression: _get_ticker_tz() args were being swapped. Improve its unit test #1793
- Refactor Ticker proxy #1711
- Add Ruff linter checks #1756
- Resolve Pandas FutureWarnings #1766
0.2.33
------
Cookie fixes:
- fix backup strategy #1759
- fix Ticker(ISIN) #1760
0.2.32
------
Add cookie & crumb to requests #1657
0.2.31
------
- Fix TZ cache exception blocking import #1705 #1709
- Fix merging pre-market events with intraday prices #1703
0.2.30
------
- Fix OperationalError #1698
0.2.29
------
- Fix pandas warning when retrieving quotes. #1672
- Replace sqlite3 with peewee for 100% thread-safety #1675
- Fix merging events with intraday prices #1684
- Fix error when calling enable_debug_mode twice #1687
- Price repair fixes #1688
0.2.28
------
- Fix TypeError: 'FastInfo' object is not callable #1636
- Improve & fix price repair #1633 #1660
- option_chain() also return underlying data #1606
0.2.27
------
Bug fixes:
- fix merging 1d-prices with out-of-range divs/splits #1635
- fix multithread error 'tz already in cache' #1648
0.2.26
------
Proxy improvements
- bug fixes #1371
- security fix #1625
0.2.25
------
Fix single ISIN as ticker #1611
Fix 'Only 100 years allowed' error #1576
0.2.24
------
Fix info[] missing values #1603
0.2.23
------
Fix 'Unauthorized' error #1595
0.2.22
------
Fix unhandled 'sqlite3.DatabaseError' #1574
0.2.21
------
Fix financials tables #1568
Price repair update: fix Yahoo messing up dividend and split adjustments #1543
Fix logging behaviour #1562
Fix merge future div/split into prices #1567
0.2.20
------
Switch to `logging` module #1493 #1522 #1541
Price history:
- optimise #1514
- fixes #1523
- fix TZ-cache corruption #1528
0.2.18
------
Fix 'fast_info' error '_np not found' #1496
Fix bug in timezone cache #1498
0.2.17
------
Fix prices error with Pandas 2.0 #1488
0.2.16
------
Fix 'fast_info deprecated' msg appearing at Ticker() init
0.2.15
------
Restore missing Ticker.info keys #1480
0.2.14
------
Fix Ticker.info dict by fetching from API #1461
0.2.13
------
Price bug fixes:
- fetch big-interval with Capital Gains #1455
- merging dividends & splits with prices #1452
0.2.12
------
Disable annoying 'backup decrypt' msg
0.2.11
------
Fix history_metadata accesses for unusual symbols #1411
0.2.10
------
General
- allow using sqlite3 < 3.8.2 #1380
- add another backup decrypt option #1379
Prices
- restore original download() timezone handling #1385
- fix & improve price repair #1289 2a2928b 86d6acc
- drop intraday intervals if in post-market but prepost=False #1311
Info
- fast_info improvements:
- add camelCase keys, add dict functions values() & items() #1368
- fix fast_info["previousClose"] #1383
- catch TypeError Exception #1397
0.2.9
-----
- Fix fast_info bugs #1362
0.2.7
-----
- Fix Yahoo decryption, smarter this time #1353
- Rename basic_info -> fast_info #1354
0.2.6
-----
- Fix Ticker.basic_info lazy-loading #1342

15
CODE_OF_CONDUCT.md Normal file
View File

@@ -0,0 +1,15 @@
# Code of Conduct
## Submitting a new issue
* Search through existing Issues and Discussions, in case your issue already exists and a solution is being developed.
* Ensure you read & follow the template form.
* Consider you may be the best person to investigate and fix.
## Contributing to an existing Issue
* Read the entire thread.
* Ensure your comment is contributing something new/useful. Remember you can simply react to other comments.
* Be concise:
- use the formatting options
- if replying to a big comment, instead of quoting it, link to it

209
README.md
View File

@@ -42,12 +42,25 @@ Yahoo! finance API is intended for personal use only.**
---
## What's new in version 0.2
## Installation
- Optimised web scraping
- All 3 financials tables now match website so expect keys to change. If you really want old tables, use [`Ticker.get_[income_stmt|balance_sheet|cashflow](legacy=True, ...)`](https://github.com/ranaroussi/yfinance/blob/85783da515761a145411d742c2a8a3c1517264b0/yfinance/base.py#L968)
- price data improvements: fix bug NaN rows with dividend; new repair feature for missing or 100x prices `download(repair=True)`; new attribute `Ticker.history_metadata`
[See release notes for full list of changes](https://github.com/ranaroussi/yfinance/releases/tag/0.2.1)
Install `yfinance` using `pip`:
``` {.sourceCode .bash}
$ pip install yfinance --upgrade --no-cache-dir
```
[With Conda](https://anaconda.org/ranaroussi/yfinance).
To install with optional dependencies, replace `optional` with: `nospam` for [caching-requests](#smarter-scraping), `repair` for [price repair](https://github.com/ranaroussi/yfinance/wiki/Price-repair), or `nospam,repair` for both:
``` {.sourceCode .bash}
$ pip install "yfinance[optional]"
```
[Required dependencies](./requirements.txt) , [all dependencies](./setup.py#L62).
---
## Quick Start
@@ -60,33 +73,23 @@ import yfinance as yf
msft = yf.Ticker("MSFT")
# fast access to subset of stock info
msft.basic_info
# slow access to all stock info
# get all stock info
msft.info
# get historical market data
hist = msft.history(period="max")
hist = msft.history(period="1mo")
# show meta information about the history (requires history() to be called first)
msft.history_metadata
# show actions (dividends, splits, capital gains)
msft.actions
# show dividends
msft.dividends
# show splits
msft.splits
# show capital gains (for mutual funds & etfs)
msft.capital_gains
msft.capital_gains # only for mutual funds & etfs
# show share count
msft.shares
msft.get_shares_full()
msft.get_shares_full(start="2022-01-01", end=None)
# show financials:
# - income statement
@@ -100,33 +103,18 @@ msft.cashflow
msft.quarterly_cashflow
# see `Ticker.get_income_stmt()` for more options
# show major holders
# show holders
msft.major_holders
# show institutional holders
msft.institutional_holders
# show mutualfund holders
msft.mutualfund_holders
msft.insider_transactions
msft.insider_purchases
msft.insider_roster_holders
# show earnings
msft.earnings
msft.quarterly_earnings
# show sustainability
msft.sustainability
# show analysts recommendations
# show recommendations
msft.recommendations
msft.recommendations_summary
# show analysts other work
msft.analyst_price_target
msft.revenue_forecasts
msft.earnings_forecasts
msft.earnings_trend
# show next event (earnings, etc)
msft.calendar
msft.upgrades_downgrades
# Show future and historic earnings dates, returns at most next 4 quarters and last 8 quarters by default.
# Note: If more are needed use msft.get_earnings_dates(limit=XX) with increased limit argument.
@@ -165,18 +153,7 @@ msft.option_chain(..., proxy="PROXY_SERVER")
...
```
To use a custom `requests` session (for example to cache calls to the
API or customize the `User-agent` header), pass a `session=` argument to
the Ticker constructor.
```python
import requests_cache
session = requests_cache.CachedSession('yfinance.cache')
session.headers['User-agent'] = 'my-program/1.0'
ticker = yf.Ticker('msft', session=session)
# The scraped response will be stored in the cache
ticker.actions
```
### Multiple tickers
To initialize multiple `Ticker` objects, use
@@ -191,69 +168,49 @@ tickers.tickers['AAPL'].history(period="1mo")
tickers.tickers['GOOG'].actions
```
### Fetching data for multiple tickers
To download price history into one table:
```python
import yfinance as yf
data = yf.download("SPY AAPL", start="2017-01-01", end="2017-04-30")
data = yf.download("SPY AAPL", period="1mo")
```
I've also added some options to make life easier :)
#### `yf.download()` and `Ticker.history()` have many options for configuring fetching and processing. [Review the Wiki](https://github.com/ranaroussi/yfinance/wiki) for more options and detail.
### Logging
`yfinance` now uses the `logging` module to handle messages, default behaviour is only print errors. If debugging, use `yf.enable_debug_mode()` to switch logging to debug with custom formatting.
### Smarter scraping
Install the `nospam` packages for smarter scraping using `pip` (see [Installation](#installation)). These packages help cache calls such that Yahoo is not spammed with requests.
To use a custom `requests` session, pass a `session=` argument to
the Ticker constructor. This allows for caching calls to the API as well as a custom way to modify requests via the `User-agent` header.
```python
data = yf.download( # or pdr.get_data_yahoo(...
# tickers list or string as well
tickers = "SPY AAPL MSFT",
# use "period" instead of start/end
# valid periods: 1d,5d,1mo,3mo,6mo,1y,2y,5y,10y,ytd,max
# (optional, default is '1mo')
period = "ytd",
# fetch data by interval (including intraday if period < 60 days)
# valid intervals: 1m,2m,5m,15m,30m,60m,90m,1h,1d,5d,1wk,1mo,3mo
# (optional, default is '1d')
interval = "5d",
# Whether to ignore timezone when aligning ticker data from
# different timezones. Default is False.
ignore_tz = False,
# group by ticker (to access via data['SPY'])
# (optional, default is 'column')
group_by = 'ticker',
# adjust all OHLC automatically
# (optional, default is False)
auto_adjust = True,
# attempt repair of missing data or currency mixups e.g. $/cents
repair = False,
# download pre/post regular market hours data
# (optional, default is False)
prepost = True,
# use threads for mass downloading? (True/False/Integer)
# (optional, default is True)
threads = True,
# proxy URL scheme use use when downloading?
# (optional, default is None)
proxy = None
)
import requests_cache
session = requests_cache.CachedSession('yfinance.cache')
session.headers['User-agent'] = 'my-program/1.0'
ticker = yf.Ticker('msft', session=session)
# The scraped response will be stored in the cache
ticker.actions
```
### Timezone cache store
When fetching price data, all dates are localized to stock exchange timezone.
But timezone retrieval is relatively slow, so yfinance attemps to cache them
in your users cache folder.
You can direct cache to use a different location with `set_tz_cache_location()`:
Combine `requests_cache` with rate-limiting to avoid triggering Yahoo's rate-limiter/blocker that can corrupt data.
```python
import yfinance as yf
yf.set_tz_cache_location("custom/cache/location")
...
from requests import Session
from requests_cache import CacheMixin, SQLiteCache
from requests_ratelimiter import LimiterMixin, MemoryQueueBucket
from pyrate_limiter import Duration, RequestRate, Limiter
class CachedLimiterSession(CacheMixin, LimiterMixin, Session):
pass
session = CachedLimiterSession(
limiter=Limiter(RequestRate(2, Duration.SECOND*5)), # max 2 requests per 5 seconds
bucket_class=MemoryQueueBucket,
backend=SQLiteCache("yfinance.cache"),
)
```
### Managing Multi-Level Columns
@@ -271,9 +228,7 @@ yfinance?](https://stackoverflow.com/questions/63107801)
- How to download single or multiple tickers into a single
dataframe with single level column names and a ticker column
---
## `pandas_datareader` override
### `pandas_datareader` override
If your code uses `pandas_datareader` and you want to download data
faster, you can "hijack" `pandas_datareader.data.get_data_yahoo()`
@@ -290,37 +245,25 @@ yf.pdr_override() # <== that's all it takes :-)
data = pdr.get_data_yahoo("SPY", start="2017-01-01", end="2017-04-30")
```
---
### Persistent cache store
## Installation
To reduce Yahoo, yfinance store some data locally: timezones to localize dates, and cookie. Cache location is:
- Windows = C:/Users/\<USER\>/AppData/Local/py-yfinance
- Linux = /home/\<USER\>/.cache/py-yfinance
- MacOS = /Users/\<USER\>/Library/Caches/py-yfinance
Install `yfinance` using `pip`:
``` {.sourceCode .bash}
$ pip install yfinance --upgrade --no-cache-dir
You can direct cache to use a different location with `set_tz_cache_location()`:
```python
import yfinance as yf
yf.set_tz_cache_location("custom/cache/location")
...
```
To install `yfinance` using `conda`, see
[this](https://anaconda.org/ranaroussi/yfinance).
---
### Requirements
## Developers: want to contribute?
- [Python](https://www.python.org) \>= 2.7, 3.4+
- [Pandas](https://github.com/pydata/pandas) \>= 1.3.0
- [Numpy](http://www.numpy.org) \>= 1.16.5
- [requests](http://docs.python-requests.org/en/master) \>= 2.26
- [lxml](https://pypi.org/project/lxml) \>= 4.9.1
- [appdirs](https://pypi.org/project/appdirs) \>= 1.4.4
- [pytz](https://pypi.org/project/pytz) \>=2022.5
- [frozendict](https://pypi.org/project/frozendict) \>= 2.3.4
- [beautifulsoup4](https://pypi.org/project/beautifulsoup4) \>= 4.11.1
- [html5lib](https://pypi.org/project/html5lib) \>= 1.1
- [cryptography](https://pypi.org/project/cryptography) \>= 3.3.2
### Optional (if you want to use `pandas_datareader`)
- [pandas\_datareader](https://github.com/pydata/pandas-datareader)
\>= 0.4.0
`yfinance` relies on community to investigate bugs and contribute code. Developer guide: https://github.com/ranaroussi/yfinance/discussions/1084
---

View File

@@ -1,5 +1,5 @@
{% set name = "yfinance" %}
{% set version = "0.2.6" %}
{% set version = "0.2.38" %}
package:
name: "{{ name|lower }}"
@@ -18,7 +18,7 @@ requirements:
host:
- pandas >=1.3.0
- numpy >=1.16.5
- requests >=2.26
- requests >=2.31
- multitasking >=0.0.7
- lxml >=4.9.1
- appdirs >=1.4.4
@@ -26,15 +26,15 @@ requirements:
- frozendict >=2.3.4
- beautifulsoup4 >=4.11.1
- html5lib >=1.1
- peewee >=3.16.2
# - pycryptodome >=3.6.6
- cryptography >=3.3.2
- pip
- python
run:
- pandas >=1.3.0
- numpy >=1.16.5
- requests >=2.26
- requests >=2.31
- multitasking >=0.0.7
- lxml >=4.9.1
- appdirs >=1.4.4
@@ -42,8 +42,8 @@ requirements:
- frozendict >=2.3.4
- beautifulsoup4 >=4.11.1
- html5lib >=1.1
- peewee >=3.16.2
# - pycryptodome >=3.6.6
- cryptography >=3.3.2
- python
test:

View File

@@ -1,6 +1,6 @@
pandas>=1.3.0
numpy>=1.16.5
requests>=2.26
requests>=2.31
multitasking>=0.0.7
lxml>=4.9.1
appdirs>=1.4.4
@@ -8,4 +8,4 @@ pytz>=2022.5
frozendict>=2.3.4
beautifulsoup4>=4.11.1
html5lib>=1.1
cryptography>=3.3.2
peewee>=3.16.2

View File

@@ -39,7 +39,7 @@ setup(
'License :: OSI Approved :: Apache Software License',
# 'Development Status :: 3 - Alpha',
'Development Status :: 4 - Beta',
#'Development Status :: 5 - Production/Stable',
# 'Development Status :: 5 - Production/Stable',
'Operating System :: OS Independent',
@@ -60,12 +60,15 @@ setup(
keywords='pandas, yahoo finance, pandas datareader',
packages=find_packages(exclude=['contrib', 'docs', 'tests', 'examples']),
install_requires=['pandas>=1.3.0', 'numpy>=1.16.5',
'requests>=2.26', 'multitasking>=0.0.7',
'requests>=2.31', 'multitasking>=0.0.7',
'lxml>=4.9.1', 'appdirs>=1.4.4', 'pytz>=2022.5',
'frozendict>=2.3.4',
# 'pycryptodome>=3.6.6',
'cryptography>=3.3.2',
'frozendict>=2.3.4', 'peewee>=3.16.2',
'beautifulsoup4>=4.11.1', 'html5lib>=1.1'],
extras_require={
'nospam': ['requests_cache>=1.0', 'requests_ratelimiter>=0.3.1'],
'repair': ['scipy>=1.6.3'],
},
# Note: Pandas.read_html() needs html5lib & beautifulsoup4
entry_points={
'console_scripts': [
'sample=sample:main',

View File

@@ -1,70 +0,0 @@
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
#
# yfinance - market data downloader
# https://github.com/ranaroussi/yfinance
"""
Sanity check for most common library uses all working
- Stock: Microsoft
- ETF: Russell 2000 Growth
- Mutual fund: Vanguard 500 Index fund
- Index: S&P500
- Currency BTC-USD
"""
import yfinance as yf
import unittest
symbols = ['MSFT', 'IWO', 'VFINX', '^GSPC', 'BTC-USD']
tickers = [yf.Ticker(symbol) for symbol in symbols]
class TestTicker(unittest.TestCase):
def test_info_history(self):
for ticker in tickers:
# always should have info and history for valid symbols
assert(ticker.info is not None and ticker.info != {})
history = ticker.history(period="max")
assert(history.empty is False and history is not None)
def test_attributes(self):
for ticker in tickers:
ticker.isin
ticker.major_holders
ticker.institutional_holders
ticker.mutualfund_holders
ticker.dividends
ticker.splits
ticker.actions
ticker.shares
ticker.info
ticker.calendar
ticker.recommendations
ticker.earnings
ticker.quarterly_earnings
ticker.income_stmt
ticker.quarterly_income_stmt
ticker.balance_sheet
ticker.quarterly_balance_sheet
ticker.cashflow
ticker.quarterly_cashflow
ticker.recommendations_summary
ticker.analyst_price_target
ticker.revenue_forecasts
ticker.sustainability
ticker.options
ticker.news
ticker.earnings_trend
ticker.earnings_dates
ticker.earnings_forecasts
def test_holders(self):
for ticker in tickers:
assert(ticker.info is not None and ticker.info != {})
assert(ticker.major_holders is not None)
assert(ticker.institutional_holders is not None)
if __name__ == '__main__':
unittest.main()

View File

@@ -1,9 +1,45 @@
# -*- coding: utf-8 -*-
import appdirs as _ad
import datetime as _dt
import sys
import os
import yfinance
from requests import Session
from requests_cache import CacheMixin, SQLiteCache
from requests_ratelimiter import LimiterMixin, MemoryQueueBucket
from pyrate_limiter import Duration, RequestRate, Limiter
_parent_dp = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
_src_dp = _parent_dp
sys.path.insert(0, _src_dp)
import yfinance
# Optional: see the exact requests that are made during tests:
# import logging
# logging.basicConfig(level=logging.DEBUG)
# Use adjacent cache folder for testing, delete if already exists and older than today
testing_cache_dirpath = os.path.join(_ad.user_cache_dir(), "py-yfinance-testing")
yfinance.set_tz_cache_location(testing_cache_dirpath)
if os.path.isdir(testing_cache_dirpath):
mtime = _dt.datetime.fromtimestamp(os.path.getmtime(testing_cache_dirpath))
if mtime.date() < _dt.date.today():
import shutil
shutil.rmtree(testing_cache_dirpath)
# Setup a session to rate-limit and cache persistently:
class CachedLimiterSession(CacheMixin, LimiterMixin, Session):
pass
history_rate = RequestRate(1, Duration.SECOND*2)
limiter = Limiter(history_rate)
cache_fp = os.path.join(testing_cache_dirpath, "unittests-cache")
session_gbl = CachedLimiterSession(
limiter=limiter,
bucket_class=MemoryQueueBucket,
backend=SQLiteCache(cache_fp, expire_after=_dt.timedelta(hours=1)),
)
# Use this instead if only want rate-limiting:
# from requests_ratelimiter import LimiterSession
# session_gbl = LimiterSession(limiter=limiter)

View File

@@ -0,0 +1,23 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-04-14 00:00:00+09:00,4126,4130,4055,4129,4129,7459400,0,0
2023-04-13 00:00:00+09:00,4064,4099,4026,4081,4081,5160200,0,0
2023-04-12 00:00:00+09:00,3968,4084,3966,4064,4064,6372000,0,0
2023-04-11 00:00:00+09:00,3990,4019,3954,3960,3960,6476500,0,0
2023-04-10 00:00:00+09:00,3996,4009,3949,3964,3964,3485200,0,0
2023-04-07 00:00:00+09:00,3897,3975,3892,3953,3953,4554700,0,0
2023-04-06 00:00:00+09:00,4002,4004,3920,3942,3942,8615200,0,0
2023-04-05 00:00:00+09:00,4150,4150,4080,4088,4088,6063700,0,0
2023-04-04 00:00:00+09:00,4245,4245,4144,4155,4155,6780600,0,0
2023-04-03 00:00:00+09:00,4250,4259,4162,4182,4182,7076800,0,0
2023-03-31 00:00:00+09:00,4229,4299,4209,4275,4275,9608400,0,0
2023-03-30 00:00:00+09:00,4257,4268,4119,4161,4161,5535200,55,5
2023-03-29 00:00:00+09:00,4146,4211,4146,4206,4151,6514500,0,0
2023-03-28 00:00:00+09:00,4200,4207,4124,4142,4087.837109375,4505500,0,0
2023-03-27 00:00:00+09:00,4196,4204,4151,4192,4137.183203125,5959500,0,0
2023-03-24 00:00:00+09:00,4130,4187,4123,4177,4122.379296875,8961500,0,0
2023-03-23 00:00:00+09:00,4056,4106,4039,4086,4032.569140625,5480000,0,0
2023-03-22 00:00:00+09:00,4066,4128,4057,4122,4068.0984375,8741500,0,0
2023-03-20 00:00:00+09:00,4000,4027,3980,3980,3927.95546875,7006500,0,0
2023-03-17 00:00:00+09:00,4018,4055,4016,4031,3978.28828125,6961500,0,0
2023-03-16 00:00:00+09:00,3976,4045,3972,4035,3982.236328125,5019000,0,0
2023-03-15 00:00:00+09:00,4034,4050,4003,4041,3988.1578125,6122000,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-04-14 00:00:00+09:00 4126 4130 4055 4129 4129 7459400 0 0
3 2023-04-13 00:00:00+09:00 4064 4099 4026 4081 4081 5160200 0 0
4 2023-04-12 00:00:00+09:00 3968 4084 3966 4064 4064 6372000 0 0
5 2023-04-11 00:00:00+09:00 3990 4019 3954 3960 3960 6476500 0 0
6 2023-04-10 00:00:00+09:00 3996 4009 3949 3964 3964 3485200 0 0
7 2023-04-07 00:00:00+09:00 3897 3975 3892 3953 3953 4554700 0 0
8 2023-04-06 00:00:00+09:00 4002 4004 3920 3942 3942 8615200 0 0
9 2023-04-05 00:00:00+09:00 4150 4150 4080 4088 4088 6063700 0 0
10 2023-04-04 00:00:00+09:00 4245 4245 4144 4155 4155 6780600 0 0
11 2023-04-03 00:00:00+09:00 4250 4259 4162 4182 4182 7076800 0 0
12 2023-03-31 00:00:00+09:00 4229 4299 4209 4275 4275 9608400 0 0
13 2023-03-30 00:00:00+09:00 4257 4268 4119 4161 4161 5535200 55 5
14 2023-03-29 00:00:00+09:00 4146 4211 4146 4206 4151 6514500 0 0
15 2023-03-28 00:00:00+09:00 4200 4207 4124 4142 4087.837109375 4505500 0 0
16 2023-03-27 00:00:00+09:00 4196 4204 4151 4192 4137.183203125 5959500 0 0
17 2023-03-24 00:00:00+09:00 4130 4187 4123 4177 4122.379296875 8961500 0 0
18 2023-03-23 00:00:00+09:00 4056 4106 4039 4086 4032.569140625 5480000 0 0
19 2023-03-22 00:00:00+09:00 4066 4128 4057 4122 4068.0984375 8741500 0 0
20 2023-03-20 00:00:00+09:00 4000 4027 3980 3980 3927.95546875 7006500 0 0
21 2023-03-17 00:00:00+09:00 4018 4055 4016 4031 3978.28828125 6961500 0 0
22 2023-03-16 00:00:00+09:00 3976 4045 3972 4035 3982.236328125 5019000 0 0
23 2023-03-15 00:00:00+09:00 4034 4050 4003 4041 3988.1578125 6122000 0 0

View File

@@ -0,0 +1,23 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-04-14 00:00:00+09:00,4126,4130,4055,4129,4129,7459400,0,0
2023-04-13 00:00:00+09:00,4064,4099,4026,4081,4081,5160200,0,0
2023-04-12 00:00:00+09:00,3968,4084,3966,4064,4064,6372000,0,0
2023-04-11 00:00:00+09:00,3990,4019,3954,3960,3960,6476500,0,0
2023-04-10 00:00:00+09:00,3996,4009,3949,3964,3964,3485200,0,0
2023-04-07 00:00:00+09:00,3897,3975,3892,3953,3953,4554700,0,0
2023-04-06 00:00:00+09:00,4002,4004,3920,3942,3942,8615200,0,0
2023-04-05 00:00:00+09:00,4150,4150,4080,4088,4088,6063700,0,0
2023-04-04 00:00:00+09:00,4245,4245,4144,4155,4155,6780600,0,0
2023-04-03 00:00:00+09:00,4250,4259,4162,4182,4182,7076800,0,0
2023-03-31 00:00:00+09:00,4229,4299,4209,4275,4275,9608400,0,0
2023-03-30 00:00:00+09:00,4257,4268,4119,4161,4161,5535200,55,5
2023-03-29 00:00:00+09:00,4146,4211,4146,4206,4151,6514500,0,0
2023-03-28 00:00:00+09:00,21000,21035,20620,20710,20439.185546875,901100,0,0
2023-03-27 00:00:00+09:00,20980,21020,20755,20960,20685.916015625,1191900,0,0
2023-03-24 00:00:00+09:00,20650,20935,20615,20885,20611.896484375,1792300,0,0
2023-03-23 00:00:00+09:00,20280,20530,20195,20430,20162.845703125,1096000,0,0
2023-03-22 00:00:00+09:00,20330,20640,20285,20610,20340.4921875,1748300,0,0
2023-03-20 00:00:00+09:00,20000,20135,19900,19900,19639.77734375,1401300,0,0
2023-03-17 00:00:00+09:00,20090,20275,20080,20155,19891.44140625,1392300,0,0
2023-03-16 00:00:00+09:00,19880,20225,19860,20175,19911.181640625,1003800,0,0
2023-03-15 00:00:00+09:00,20170,20250,20015,20205,19940.7890625,1224400,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-04-14 00:00:00+09:00 4126 4130 4055 4129 4129 7459400 0 0
3 2023-04-13 00:00:00+09:00 4064 4099 4026 4081 4081 5160200 0 0
4 2023-04-12 00:00:00+09:00 3968 4084 3966 4064 4064 6372000 0 0
5 2023-04-11 00:00:00+09:00 3990 4019 3954 3960 3960 6476500 0 0
6 2023-04-10 00:00:00+09:00 3996 4009 3949 3964 3964 3485200 0 0
7 2023-04-07 00:00:00+09:00 3897 3975 3892 3953 3953 4554700 0 0
8 2023-04-06 00:00:00+09:00 4002 4004 3920 3942 3942 8615200 0 0
9 2023-04-05 00:00:00+09:00 4150 4150 4080 4088 4088 6063700 0 0
10 2023-04-04 00:00:00+09:00 4245 4245 4144 4155 4155 6780600 0 0
11 2023-04-03 00:00:00+09:00 4250 4259 4162 4182 4182 7076800 0 0
12 2023-03-31 00:00:00+09:00 4229 4299 4209 4275 4275 9608400 0 0
13 2023-03-30 00:00:00+09:00 4257 4268 4119 4161 4161 5535200 55 5
14 2023-03-29 00:00:00+09:00 4146 4211 4146 4206 4151 6514500 0 0
15 2023-03-28 00:00:00+09:00 21000 21035 20620 20710 20439.185546875 901100 0 0
16 2023-03-27 00:00:00+09:00 20980 21020 20755 20960 20685.916015625 1191900 0 0
17 2023-03-24 00:00:00+09:00 20650 20935 20615 20885 20611.896484375 1792300 0 0
18 2023-03-23 00:00:00+09:00 20280 20530 20195 20430 20162.845703125 1096000 0 0
19 2023-03-22 00:00:00+09:00 20330 20640 20285 20610 20340.4921875 1748300 0 0
20 2023-03-20 00:00:00+09:00 20000 20135 19900 19900 19639.77734375 1401300 0 0
21 2023-03-17 00:00:00+09:00 20090 20275 20080 20155 19891.44140625 1392300 0 0
22 2023-03-16 00:00:00+09:00 19880 20225 19860 20175 19911.181640625 1003800 0 0
23 2023-03-15 00:00:00+09:00 20170 20250 20015 20205 19940.7890625 1224400 0 0

View File

@@ -0,0 +1,6 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-30 00:00:00+02:00,19.5900001525879,19.7999992370605,19.2700004577637,19.3500003814697,18.6291382416581,196309,0,0
2023-05-31 00:00:00+02:00,19.1200008392334,19.1399993896484,18.7000007629395,18.7900009155273,18.0900009155273,156652,0,0
2023-06-02 00:00:00+02:00,18.5499992370605,19,18.5100002288818,18.8999996185303,18.8999996185303,83439,0.7,0
2023-06-05 00:00:00+02:00,18.9300003051758,19.0900001525879,18.8400001525879,19,19,153167,0,0
2023-06-06 00:00:00+02:00,18.9099998474121,18.9500007629395,18.5100002288818,18.6599998474121,18.6599998474121,104352,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-30 00:00:00+02:00 19.5900001525879 19.7999992370605 19.2700004577637 19.3500003814697 18.6291382416581 196309 0 0
3 2023-05-31 00:00:00+02:00 19.1200008392334 19.1399993896484 18.7000007629395 18.7900009155273 18.0900009155273 156652 0 0
4 2023-06-02 00:00:00+02:00 18.5499992370605 19 18.5100002288818 18.8999996185303 18.8999996185303 83439 0.7 0
5 2023-06-05 00:00:00+02:00 18.9300003051758 19.0900001525879 18.8400001525879 19 19 153167 0 0
6 2023-06-06 00:00:00+02:00 18.9099998474121 18.9500007629395 18.5100002288818 18.6599998474121 18.6599998474121 104352 0 0

View File

@@ -0,0 +1,6 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-30 00:00:00+02:00,19.59000015258789,19.799999237060547,19.270000457763672,19.350000381469727,19.350000381469727,196309,0.0,0.0
2023-05-31 00:00:00+02:00,19.1200008392334,19.139999389648438,18.700000762939453,18.790000915527344,18.790000915527344,156652,0.0,0.0
2023-06-02 00:00:00+02:00,18.549999237060547,19.0,18.510000228881836,18.899999618530273,18.899999618530273,83439,0.7,0.0
2023-06-05 00:00:00+02:00,18.93000030517578,19.09000015258789,18.84000015258789,19.0,19.0,153167,0.0,0.0
2023-06-06 00:00:00+02:00,18.90999984741211,18.950000762939453,18.510000228881836,18.65999984741211,18.65999984741211,104352,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-30 00:00:00+02:00 19.59000015258789 19.799999237060547 19.270000457763672 19.350000381469727 19.350000381469727 196309 0.0 0.0
3 2023-05-31 00:00:00+02:00 19.1200008392334 19.139999389648438 18.700000762939453 18.790000915527344 18.790000915527344 156652 0.0 0.0
4 2023-06-02 00:00:00+02:00 18.549999237060547 19.0 18.510000228881836 18.899999618530273 18.899999618530273 83439 0.7 0.0
5 2023-06-05 00:00:00+02:00 18.93000030517578 19.09000015258789 18.84000015258789 19.0 19.0 153167 0.0 0.0
6 2023-06-06 00:00:00+02:00 18.90999984741211 18.950000762939453 18.510000228881836 18.65999984741211 18.65999984741211 104352 0.0 0.0

View File

@@ -0,0 +1,24 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-06-06 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-06-01 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-31 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-30 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-27 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-26 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-25 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-24 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-23 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-20 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-19 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-18 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,532454,0,0
2022-05-17 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-16 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-13 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-12 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-11 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-10 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-09 00:00:00+01:00,0.1455,0.1455,0.1455,0.1455,0.1455,0,0,0
2022-05-06 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-05 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-04 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
2022-05-03 00:00:00+01:00,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0.145500004291534,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-06-06 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
3 2022-06-01 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
4 2022-05-31 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
5 2022-05-30 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
6 2022-05-27 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
7 2022-05-26 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
8 2022-05-25 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
9 2022-05-24 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
10 2022-05-23 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
11 2022-05-20 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
12 2022-05-19 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
13 2022-05-18 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 532454 0 0
14 2022-05-17 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
15 2022-05-16 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
16 2022-05-13 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
17 2022-05-12 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
18 2022-05-11 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
19 2022-05-10 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
20 2022-05-09 00:00:00+01:00 0.1455 0.1455 0.1455 0.1455 0.1455 0 0 0
21 2022-05-06 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
22 2022-05-05 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
23 2022-05-04 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0
24 2022-05-03 00:00:00+01:00 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0.145500004291534 0 0 0

View File

@@ -0,0 +1,24 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-06-06 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-06-01 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-31 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-30 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-27 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-26 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-25 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-24 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-23 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-20 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-19 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-18 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,532454,0.0,0.0
2022-05-17 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-16 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-13 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-12 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-11 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-10 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-09 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-06 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-05 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-04 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-03 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-06-06 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
3 2022-06-01 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
4 2022-05-31 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
5 2022-05-30 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
6 2022-05-27 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
7 2022-05-26 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
8 2022-05-25 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
9 2022-05-24 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
10 2022-05-23 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
11 2022-05-20 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
12 2022-05-19 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
13 2022-05-18 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 532454 0.0 0.0
14 2022-05-17 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
15 2022-05-16 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
16 2022-05-13 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
17 2022-05-12 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
18 2022-05-11 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
19 2022-05-10 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
20 2022-05-09 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
21 2022-05-06 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
22 2022-05-05 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
23 2022-05-04 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
24 2022-05-03 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0

View File

@@ -0,0 +1,37 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-05-30 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-05-23 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-05-16 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,532454,0,0
2022-05-09 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-05-02 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-04-25 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-04-18 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-04-11 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-04-04 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-03-28 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-03-21 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-03-14 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-03-07 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-02-28 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-02-21 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-02-14 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-02-07 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-01-31 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-01-24 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-01-17 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-01-10 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-01-03 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-12-27 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-12-20 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-12-13 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-12-06 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-11-29 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-11-22 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-11-15 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-11-08 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-11-01 00:00:00+00:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-10-25 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-10-18 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-10-11 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2021-10-04 00:00:00+01:00,14.8000,15.3400,14.4000,14.5500,14.5500,2171373,0,0
2021-09-27 00:00:00+01:00,15.6000,16.0000,14.9000,15.0500,15.0500,3860549,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-05-30 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
3 2022-05-23 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
4 2022-05-16 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 532454 0 0
5 2022-05-09 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
6 2022-05-02 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
7 2022-04-25 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
8 2022-04-18 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
9 2022-04-11 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
10 2022-04-04 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
11 2022-03-28 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
12 2022-03-21 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
13 2022-03-14 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
14 2022-03-07 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
15 2022-02-28 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
16 2022-02-21 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
17 2022-02-14 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
18 2022-02-07 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
19 2022-01-31 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
20 2022-01-24 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
21 2022-01-17 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
22 2022-01-10 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
23 2022-01-03 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
24 2021-12-27 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
25 2021-12-20 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
26 2021-12-13 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
27 2021-12-06 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
28 2021-11-29 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
29 2021-11-22 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
30 2021-11-15 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
31 2021-11-08 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
32 2021-11-01 00:00:00+00:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
33 2021-10-25 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
34 2021-10-18 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
35 2021-10-11 00:00:00+01:00 14.5500 14.5500 14.5500 14.5500 14.5500 0 0 0
36 2021-10-04 00:00:00+01:00 14.8000 15.3400 14.4000 14.5500 14.5500 2171373 0 0
37 2021-09-27 00:00:00+01:00 15.6000 16.0000 14.9000 15.0500 15.0500 3860549 0 0

View File

@@ -0,0 +1,25 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-08-15 00:00:00+01:00,27.6000,28.2000,26.2000,27.6000,27.6000,3535668,0,0
2022-08-12 00:00:00+01:00,27.3000,29.8000,26.4030,27.0000,27.0000,7223353,0,0
2022-08-11 00:00:00+01:00,26.0000,29.8000,24.2000,27.1000,27.1000,12887933,0,0
2022-08-10 00:00:00+01:00,25.0000,29.2000,22.5000,25.0000,25.0000,26572680,0,0
2022-08-09 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-08 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-05 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-04 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-03 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-02 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-08-01 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-29 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-28 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-27 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-26 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-25 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-22 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-21 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-20 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-19 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-18 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-15 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-14 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0
2022-07-13 00:00:00+01:00,14.5500,14.5500,14.5500,14.5500,14.5500,0,0,0

View File

@@ -0,0 +1,37 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-05-30 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-23 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-05-16 00:00:00+01:00,14.550000190734863,14.550000190734863,0.14550000429153442,0.14550000429153442,0.14550000429153442,532454,0.0,0.0
2022-05-09 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-05-02 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-04-25 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-04-18 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-04-11 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-04-04 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-03-28 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-03-21 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-03-14 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-03-07 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-02-28 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-02-21 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-02-14 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-02-07 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-01-31 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-01-24 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-01-17 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-01-10 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-01-03 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-12-27 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-12-20 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-12-13 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-12-06 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-11-29 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-11-22 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-11-15 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-11-08 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-11-01 00:00:00+00:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-10-25 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-10-18 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-10-11 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2021-10-04 00:00:00+01:00,14.800000190734863,15.34000015258789,0.14399999380111694,0.14550000429153442,0.14550000429153442,2171373,0.0,0.0
2021-09-27 00:00:00+01:00,15.600000381469727,16.0,14.899999618530273,15.050000190734863,15.050000190734863,3860549,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-05-30 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
3 2022-05-23 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
4 2022-05-16 00:00:00+01:00 14.550000190734863 14.550000190734863 0.14550000429153442 0.14550000429153442 0.14550000429153442 532454 0.0 0.0
5 2022-05-09 00:00:00+01:00 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 14.550000190734863 0 0.0 0.0
6 2022-05-02 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
7 2022-04-25 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
8 2022-04-18 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
9 2022-04-11 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
10 2022-04-04 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
11 2022-03-28 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
12 2022-03-21 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
13 2022-03-14 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
14 2022-03-07 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
15 2022-02-28 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
16 2022-02-21 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
17 2022-02-14 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
18 2022-02-07 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
19 2022-01-31 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
20 2022-01-24 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
21 2022-01-17 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
22 2022-01-10 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
23 2022-01-03 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
24 2021-12-27 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
25 2021-12-20 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
26 2021-12-13 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
27 2021-12-06 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
28 2021-11-29 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
29 2021-11-22 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
30 2021-11-15 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
31 2021-11-08 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
32 2021-11-01 00:00:00+00:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
33 2021-10-25 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
34 2021-10-18 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
35 2021-10-11 00:00:00+01:00 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0.14550000429153442 0 0.0 0.0
36 2021-10-04 00:00:00+01:00 14.800000190734863 15.34000015258789 0.14399999380111694 0.14550000429153442 0.14550000429153442 2171373 0.0 0.0
37 2021-09-27 00:00:00+01:00 15.600000381469727 16.0 14.899999618530273 15.050000190734863 15.050000190734863 3860549 0.0 0.0

View File

@@ -0,0 +1,25 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-08-15 00:00:00+01:00,27.600000381469727,28.200000762939453,26.200000762939453,27.600000381469727,27.600000381469727,3535668,0.0,0.0
2022-08-12 00:00:00+01:00,27.299999237060547,29.799999237060547,26.402999877929688,27.0,27.0,7223353,0.0,0.0
2022-08-11 00:00:00+01:00,26.0,29.799999237060547,24.200000762939453,27.100000381469727,27.100000381469727,12887933,0.0,0.0
2022-08-10 00:00:00+01:00,25.0,29.200000762939453,22.5,25.0,25.0,26572680,0.0,0.0
2022-08-09 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-08 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-05 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-04 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-03 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-02 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-08-01 00:00:00+01:00,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,14.550000190734863,0,0.0,0.0
2022-07-29 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-28 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-27 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-26 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-25 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-22 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-21 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-20 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-19 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-18 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-15 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-14 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0
2022-07-13 00:00:00+01:00,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0.14550000429153442,0,0.0,0.0

View File

@@ -0,0 +1,30 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-04-20 00:00:00+02:00,3,3,2,3,3,2076,0,0
2023-04-21 00:00:00+02:00,3,3,2,3,3,2136,0,0
2023-04-24 00:00:00+02:00,3,3,1,1,1,77147,0,0
2023-04-25 00:00:00+02:00,1,2,1,2,2,9625,0,0
2023-04-26 00:00:00+02:00,2,2,1,2,2,5028,0,0
2023-04-27 00:00:00+02:00,2,2,1,1,1,3235,0,0
2023-04-28 00:00:00+02:00,2,2,1,2,2,10944,0,0
2023-05-02 00:00:00+02:00,2,2,2,2,2,12220,0,0
2023-05-03 00:00:00+02:00,2,2,2,2,2,4683,0,0
2023-05-04 00:00:00+02:00,2,2,1,2,2,3368,0,0
2023-05-05 00:00:00+02:00,2,2,1,2,2,26069,0,0
2023-05-08 00:00:00+02:00,1,2,1,1,1,70540,0,0
2023-05-09 00:00:00+02:00,1,2,1,1,1,14228,0,0
2023-05-10 00:00:00+02:00,1.08000004291534,1.39999997615814,0.879999995231628,1,1,81012,0,0.0001
2023-05-11 00:00:00+02:00,1.03999996185303,1.03999996185303,0.850000023841858,1,1,40254,0,0
2023-05-12 00:00:00+02:00,0.949999988079071,1.10000002384186,0.949999988079071,1.01999998092651,1.01999998092651,35026,0,0
2023-05-15 00:00:00+02:00,0.949999988079071,1.01999998092651,0.860000014305115,0.939999997615814,0.939999997615814,41486,0,0
2023-05-16 00:00:00+02:00,0.899999976158142,0.944000005722046,0.800000011920929,0.800000011920929,0.800000011920929,43583,0,0
2023-05-17 00:00:00+02:00,0.850000023841858,0.850000023841858,0.779999971389771,0.810000002384186,0.810000002384186,29984,0,0
2023-05-18 00:00:00+02:00,0.779999971389771,0.78600001335144,0.740000009536743,0.740000009536743,0.740000009536743,24679,0,0
2023-05-19 00:00:00+02:00,0.78600001335144,0.78600001335144,0.649999976158142,0.65200001001358,0.65200001001358,26732,0,0
2023-05-22 00:00:00+02:00,0.8299999833107,1.05999994277954,0.709999978542328,0.709999978542328,0.709999978542328,169538,0,0
2023-05-23 00:00:00+02:00,0.899999976158142,1.60800004005432,0.860000014305115,1.22000002861023,1.22000002861023,858471,0,0
2023-05-24 00:00:00+02:00,1.19400000572205,1.25999999046326,0.779999971389771,0.779999971389771,0.779999971389771,627823,0,0
2023-05-25 00:00:00+02:00,0.980000019073486,1.22000002861023,0.702000021934509,0.732999980449677,0.732999980449677,1068939,0,0
2023-05-26 00:00:00+02:00,0.660000026226044,0.72000002861023,0.602999985218048,0.611999988555908,0.611999988555908,631580,0,0
2023-05-29 00:00:00+02:00,0.620000004768372,0.75,0.578999996185303,0.600000023841858,0.600000023841858,586150,0,0
2023-05-30 00:00:00+02:00,0.610000014305115,0.634999990463257,0.497000008821487,0.497000008821487,0.497000008821487,552308,0,0
2023-05-31 00:00:00+02:00,0.458999991416931,0.469999998807907,0.374000012874603,0.379999995231628,0.379999995231628,899067,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-04-20 00:00:00+02:00 3 3 2 3 3 2076 0 0
3 2023-04-21 00:00:00+02:00 3 3 2 3 3 2136 0 0
4 2023-04-24 00:00:00+02:00 3 3 1 1 1 77147 0 0
5 2023-04-25 00:00:00+02:00 1 2 1 2 2 9625 0 0
6 2023-04-26 00:00:00+02:00 2 2 1 2 2 5028 0 0
7 2023-04-27 00:00:00+02:00 2 2 1 1 1 3235 0 0
8 2023-04-28 00:00:00+02:00 2 2 1 2 2 10944 0 0
9 2023-05-02 00:00:00+02:00 2 2 2 2 2 12220 0 0
10 2023-05-03 00:00:00+02:00 2 2 2 2 2 4683 0 0
11 2023-05-04 00:00:00+02:00 2 2 1 2 2 3368 0 0
12 2023-05-05 00:00:00+02:00 2 2 1 2 2 26069 0 0
13 2023-05-08 00:00:00+02:00 1 2 1 1 1 70540 0 0
14 2023-05-09 00:00:00+02:00 1 2 1 1 1 14228 0 0
15 2023-05-10 00:00:00+02:00 1.08000004291534 1.39999997615814 0.879999995231628 1 1 81012 0 0.0001
16 2023-05-11 00:00:00+02:00 1.03999996185303 1.03999996185303 0.850000023841858 1 1 40254 0 0
17 2023-05-12 00:00:00+02:00 0.949999988079071 1.10000002384186 0.949999988079071 1.01999998092651 1.01999998092651 35026 0 0
18 2023-05-15 00:00:00+02:00 0.949999988079071 1.01999998092651 0.860000014305115 0.939999997615814 0.939999997615814 41486 0 0
19 2023-05-16 00:00:00+02:00 0.899999976158142 0.944000005722046 0.800000011920929 0.800000011920929 0.800000011920929 43583 0 0
20 2023-05-17 00:00:00+02:00 0.850000023841858 0.850000023841858 0.779999971389771 0.810000002384186 0.810000002384186 29984 0 0
21 2023-05-18 00:00:00+02:00 0.779999971389771 0.78600001335144 0.740000009536743 0.740000009536743 0.740000009536743 24679 0 0
22 2023-05-19 00:00:00+02:00 0.78600001335144 0.78600001335144 0.649999976158142 0.65200001001358 0.65200001001358 26732 0 0
23 2023-05-22 00:00:00+02:00 0.8299999833107 1.05999994277954 0.709999978542328 0.709999978542328 0.709999978542328 169538 0 0
24 2023-05-23 00:00:00+02:00 0.899999976158142 1.60800004005432 0.860000014305115 1.22000002861023 1.22000002861023 858471 0 0
25 2023-05-24 00:00:00+02:00 1.19400000572205 1.25999999046326 0.779999971389771 0.779999971389771 0.779999971389771 627823 0 0
26 2023-05-25 00:00:00+02:00 0.980000019073486 1.22000002861023 0.702000021934509 0.732999980449677 0.732999980449677 1068939 0 0
27 2023-05-26 00:00:00+02:00 0.660000026226044 0.72000002861023 0.602999985218048 0.611999988555908 0.611999988555908 631580 0 0
28 2023-05-29 00:00:00+02:00 0.620000004768372 0.75 0.578999996185303 0.600000023841858 0.600000023841858 586150 0 0
29 2023-05-30 00:00:00+02:00 0.610000014305115 0.634999990463257 0.497000008821487 0.497000008821487 0.497000008821487 552308 0 0
30 2023-05-31 00:00:00+02:00 0.458999991416931 0.469999998807907 0.374000012874603 0.379999995231628 0.379999995231628 899067 0 0

View File

@@ -0,0 +1,30 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-04-20 00:00:00+02:00,3.0,3.0,2.0,3.0,3.0,2076,0.0,0.0
2023-04-21 00:00:00+02:00,3.0,3.0,2.0,3.0,3.0,2136,0.0,0.0
2023-04-24 00:00:00+02:00,3.0,3.0,1.0,1.0,1.0,77147,0.0,0.0
2023-04-25 00:00:00+02:00,1.0,2.0,1.0,2.0,2.0,9625,0.0,0.0
2023-04-26 00:00:00+02:00,2.0,2.0,1.0,2.0,2.0,5028,0.0,0.0
2023-04-27 00:00:00+02:00,2.0,2.0,1.0,1.0,1.0,3235,0.0,0.0
2023-04-28 00:00:00+02:00,2.0,2.0,1.0,2.0,2.0,10944,0.0,0.0
2023-05-02 00:00:00+02:00,2.0,2.0,2.0,2.0,2.0,12220,0.0,0.0
2023-05-03 00:00:00+02:00,2.0,2.0,2.0,2.0,2.0,4683,0.0,0.0
2023-05-04 00:00:00+02:00,2.0,2.0,1.0,2.0,2.0,3368,0.0,0.0
2023-05-05 00:00:00+02:00,2.0,2.0,1.0,2.0,2.0,26069,0.0,0.0
2023-05-08 00:00:00+02:00,9.999999747378752e-05,0.00019999999494757503,9.999999747378752e-05,9.999999747378752e-05,9.999999747378752e-05,705399568,0.0,0.0
2023-05-09 00:00:00+02:00,1.0,2.0,1.0,1.0,1.0,14228,0.0,0.0
2023-05-10 00:00:00+02:00,1.0800000429153442,1.399999976158142,0.8799999952316284,1.0,1.0,81012,0.0,0.0001
2023-05-11 00:00:00+02:00,1.0399999618530273,1.0399999618530273,0.8500000238418579,1.0,1.0,40254,0.0,0.0
2023-05-12 00:00:00+02:00,0.949999988079071,1.100000023841858,0.949999988079071,1.0199999809265137,1.0199999809265137,35026,0.0,0.0
2023-05-15 00:00:00+02:00,0.949999988079071,1.0199999809265137,0.8600000143051147,0.9399999976158142,0.9399999976158142,41486,0.0,0.0
2023-05-16 00:00:00+02:00,0.8999999761581421,0.9440000057220459,0.800000011920929,0.800000011920929,0.800000011920929,43583,0.0,0.0
2023-05-17 00:00:00+02:00,0.8500000238418579,0.8500000238418579,0.7799999713897705,0.8100000023841858,0.8100000023841858,29984,0.0,0.0
2023-05-18 00:00:00+02:00,0.7799999713897705,0.7860000133514404,0.7400000095367432,0.7400000095367432,0.7400000095367432,24679,0.0,0.0
2023-05-19 00:00:00+02:00,0.7860000133514404,0.7860000133514404,0.6499999761581421,0.6520000100135803,0.6520000100135803,26732,0.0,0.0
2023-05-22 00:00:00+02:00,0.8299999833106995,1.059999942779541,0.7099999785423279,0.7099999785423279,0.7099999785423279,169538,0.0,0.0
2023-05-23 00:00:00+02:00,0.8999999761581421,1.6080000400543213,0.8600000143051147,1.2200000286102295,1.2200000286102295,858471,0.0,0.0
2023-05-24 00:00:00+02:00,1.194000005722046,1.2599999904632568,0.7799999713897705,0.7799999713897705,0.7799999713897705,627823,0.0,0.0
2023-05-25 00:00:00+02:00,0.9800000190734863,1.2200000286102295,0.7020000219345093,0.7329999804496765,0.7329999804496765,1068939,0.0,0.0
2023-05-26 00:00:00+02:00,0.6600000262260437,0.7200000286102295,0.6029999852180481,0.6119999885559082,0.6119999885559082,631580,0.0,0.0
2023-05-29 00:00:00+02:00,0.6200000047683716,0.75,0.5789999961853027,0.6000000238418579,0.6000000238418579,586150,0.0,0.0
2023-05-30 00:00:00+02:00,0.6100000143051147,0.6349999904632568,0.4970000088214874,0.4970000088214874,0.4970000088214874,552308,0.0,0.0
2023-05-31 00:00:00+02:00,0.45899999141693115,0.4699999988079071,0.37400001287460327,0.3799999952316284,0.3799999952316284,899067,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-04-20 00:00:00+02:00 3.0 3.0 2.0 3.0 3.0 2076 0.0 0.0
3 2023-04-21 00:00:00+02:00 3.0 3.0 2.0 3.0 3.0 2136 0.0 0.0
4 2023-04-24 00:00:00+02:00 3.0 3.0 1.0 1.0 1.0 77147 0.0 0.0
5 2023-04-25 00:00:00+02:00 1.0 2.0 1.0 2.0 2.0 9625 0.0 0.0
6 2023-04-26 00:00:00+02:00 2.0 2.0 1.0 2.0 2.0 5028 0.0 0.0
7 2023-04-27 00:00:00+02:00 2.0 2.0 1.0 1.0 1.0 3235 0.0 0.0
8 2023-04-28 00:00:00+02:00 2.0 2.0 1.0 2.0 2.0 10944 0.0 0.0
9 2023-05-02 00:00:00+02:00 2.0 2.0 2.0 2.0 2.0 12220 0.0 0.0
10 2023-05-03 00:00:00+02:00 2.0 2.0 2.0 2.0 2.0 4683 0.0 0.0
11 2023-05-04 00:00:00+02:00 2.0 2.0 1.0 2.0 2.0 3368 0.0 0.0
12 2023-05-05 00:00:00+02:00 2.0 2.0 1.0 2.0 2.0 26069 0.0 0.0
13 2023-05-08 00:00:00+02:00 9.999999747378752e-05 0.00019999999494757503 9.999999747378752e-05 9.999999747378752e-05 9.999999747378752e-05 705399568 0.0 0.0
14 2023-05-09 00:00:00+02:00 1.0 2.0 1.0 1.0 1.0 14228 0.0 0.0
15 2023-05-10 00:00:00+02:00 1.0800000429153442 1.399999976158142 0.8799999952316284 1.0 1.0 81012 0.0 0.0001
16 2023-05-11 00:00:00+02:00 1.0399999618530273 1.0399999618530273 0.8500000238418579 1.0 1.0 40254 0.0 0.0
17 2023-05-12 00:00:00+02:00 0.949999988079071 1.100000023841858 0.949999988079071 1.0199999809265137 1.0199999809265137 35026 0.0 0.0
18 2023-05-15 00:00:00+02:00 0.949999988079071 1.0199999809265137 0.8600000143051147 0.9399999976158142 0.9399999976158142 41486 0.0 0.0
19 2023-05-16 00:00:00+02:00 0.8999999761581421 0.9440000057220459 0.800000011920929 0.800000011920929 0.800000011920929 43583 0.0 0.0
20 2023-05-17 00:00:00+02:00 0.8500000238418579 0.8500000238418579 0.7799999713897705 0.8100000023841858 0.8100000023841858 29984 0.0 0.0
21 2023-05-18 00:00:00+02:00 0.7799999713897705 0.7860000133514404 0.7400000095367432 0.7400000095367432 0.7400000095367432 24679 0.0 0.0
22 2023-05-19 00:00:00+02:00 0.7860000133514404 0.7860000133514404 0.6499999761581421 0.6520000100135803 0.6520000100135803 26732 0.0 0.0
23 2023-05-22 00:00:00+02:00 0.8299999833106995 1.059999942779541 0.7099999785423279 0.7099999785423279 0.7099999785423279 169538 0.0 0.0
24 2023-05-23 00:00:00+02:00 0.8999999761581421 1.6080000400543213 0.8600000143051147 1.2200000286102295 1.2200000286102295 858471 0.0 0.0
25 2023-05-24 00:00:00+02:00 1.194000005722046 1.2599999904632568 0.7799999713897705 0.7799999713897705 0.7799999713897705 627823 0.0 0.0
26 2023-05-25 00:00:00+02:00 0.9800000190734863 1.2200000286102295 0.7020000219345093 0.7329999804496765 0.7329999804496765 1068939 0.0 0.0
27 2023-05-26 00:00:00+02:00 0.6600000262260437 0.7200000286102295 0.6029999852180481 0.6119999885559082 0.6119999885559082 631580 0.0 0.0
28 2023-05-29 00:00:00+02:00 0.6200000047683716 0.75 0.5789999961853027 0.6000000238418579 0.6000000238418579 586150 0.0 0.0
29 2023-05-30 00:00:00+02:00 0.6100000143051147 0.6349999904632568 0.4970000088214874 0.4970000088214874 0.4970000088214874 552308 0.0 0.0
30 2023-05-31 00:00:00+02:00 0.45899999141693115 0.4699999988079071 0.37400001287460327 0.3799999952316284 0.3799999952316284 899067 0.0 0.0

View File

@@ -0,0 +1,85 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2021-12-13 00:00:00+00:00,393.999975585938,406.6,391.4,402.899916992188,291.232287597656,62714764.4736842,0,0
2021-12-20 00:00:00+00:00,393.999975585938,412.199990234375,392.502983398438,409.899997558594,296.292243652344,46596651.3157895,0,0
2021-12-27 00:00:00+00:00,409.899997558594,416.550971679688,408.387001953125,410.4,296.653642578125,10818482.8947368,0,0
2022-01-03 00:00:00+00:00,410.4,432.199995117188,410.4,432.099985351563,312.339265136719,44427327.6315789,0,0
2022-01-10 00:00:00+00:00,431.3,439.199982910156,429.099970703125,436.099912109375,315.230618896484,29091400,0,0
2022-01-17 00:00:00+00:00,437.999912109375,445.199965820313,426.999997558594,431.999975585938,312.267017822266,43787351.3157895,0,0
2022-01-24 00:00:00+00:00,430.099975585938,440.999973144531,420.999968261719,433.499982910156,313.351237792969,58487296.0526316,0,0
2022-01-31 00:00:00+00:00,436.199968261719,443.049987792969,432.099985351563,435.199916992188,314.580045166016,43335806.5789474,0,0
2022-02-07 00:00:00+00:00,437.899995117188,448.799992675781,436.051994628906,444.39998046875,321.230207519531,39644061.8421053,0,0
2022-02-14 00:00:00+00:00,437.699975585938,441.999978027344,426.699968261719,432.199995117188,312.411558837891,49972693.4210526,0,0
2022-02-21 00:00:00+00:00,435.499992675781,438.476999511719,408.29998046875,423.399970703125,306.050571289063,65719596.0526316,0,0
2022-02-28 00:00:00+00:00,415.099995117188,427.999909667969,386.199932861328,386.799945068359,279.594578857422,94057936.8421053,4.1875,0
2022-03-07 00:00:00+00:00,374.999952392578,417.299978027344,361.101981201172,409.599968261719,298.389248046875,71269101.3157895,0,0
2022-03-14 00:00:00+00:00,413.099985351563,426.699968261719,408.899992675781,422.399965820313,307.713929443359,55431927.6315789,0,0
2022-03-21 00:00:00+00:00,422.699995117188,442.7,422.399965820313,437.799985351563,318.932696533203,39896352.6315789,0,0
2022-03-28 00:00:00+01:00,442.49998046875,460.999978027344,440.097983398438,444.6,323.886403808594,56413515.7894737,0,0
2022-04-04 00:00:00+01:00,439.699985351563,445.399985351563,421.999973144531,425.799973144531,310.190817871094,49415836.8421053,19.342106,0
2022-04-11 00:00:00+01:00,425.39998046875,435.599909667969,420.799995117188,434.299968261719,327.211427001953,29875081.5789474,0,0
2022-04-18 00:00:00+01:00,434.299968261719,447.799987792969,433.599992675781,437.799985351563,329.848419189453,49288272.3684211,0,0
2022-04-25 00:00:00+01:00,430.699987792969,438.799990234375,423.999982910156,433.299916992188,326.457967529297,44656776.3157895,0,0
2022-05-02 00:00:00+01:00,433.299916992188,450.999975585938,414.499982910156,414.899975585938,312.595018310547,29538167.1052632,0,0
2022-05-09 00:00:00+01:00,413.199995117188,417.449992675781,368.282923583984,408.199970703125,307.547099609375,73989611.8421053,0,0
2022-05-16 00:00:00+01:00,384,423.600006103516,384,412.100006103516,310.485473632813,81938261,101.69,0.76
2022-05-23 00:00:00+01:00,416.100006103516,442.399993896484,341.915008544922,440.899993896484,409.764678955078,45432941,0,0
2022-05-30 00:00:00+01:00,442.700012207031,444.200012207031,426.600006103516,428.700012207031,398.426239013672,37906659,0,0
2022-06-06 00:00:00+01:00,425.299987792969,434.010009765625,405.200012207031,405.399993896484,376.771606445313,40648810,0,0
2022-06-13 00:00:00+01:00,402.5,420,399.799987792969,411.200012207031,382.162048339844,74196958,0,0
2022-06-20 00:00:00+01:00,412.5,421.899993896484,398.399993896484,411.5,382.440826416016,28679717,0,0
2022-06-27 00:00:00+01:00,413.100006103516,422.399993896484,397.399993896484,401.600006103516,373.239959716797,35468994,0,0
2022-07-04 00:00:00+01:00,405.399993896484,406.600006103516,382.299987792969,401.299987792969,372.961120605469,35304748,0,0
2022-07-11 00:00:00+01:00,394.799987792969,405.850006103516,383.399993896484,396.600006103516,368.593048095703,42308459,0,0
2022-07-18 00:00:00+01:00,392.5,399.700012207031,384.799987792969,391.700012207031,364.039093017578,36656839,0,0
2022-07-25 00:00:00+01:00,392.200012207031,400.799987792969,388.700012207031,396,368.035430908203,33124660,0,0
2022-08-01 00:00:00+01:00,396.399993896484,405.5,390.415008544922,402,373.611724853516,21753121,0,0
2022-08-08 00:00:00+01:00,406.600006103516,473.700012207031,403.299987792969,467.899993896484,434.858032226563,59155709,0,0
2022-08-15 00:00:00+01:00,468.100006103516,470.5,434,437,406.140106201172,36989620,10.3,0
2022-08-22 00:00:00+01:00,436.100006103516,436.869995117188,419.299987792969,420.5,399.780303955078,36492572,0,0
2022-08-29 00:00:00+01:00,420.5,426.600006103516,408.600006103516,426.600006103516,405.579742431641,29573657,0,0
2022-09-05 00:00:00+01:00,418.5,444.4169921875,416.100006103516,443.100006103516,421.266723632813,34375126,0,0
2022-09-12 00:00:00+01:00,444.649993896484,448.899993896484,435.200012207031,440.100006103516,418.414520263672,39085960,0,0
2022-09-19 00:00:00+01:00,440.100006103516,447.200012207031,419.299987792969,422.899993896484,402.062042236328,27982081,0,0
2022-09-26 00:00:00+01:00,421.200012207031,421.200012207031,373.31201171875,388.200012207031,369.071868896484,70408935,0,0
2022-10-03 00:00:00+01:00,382.899993896484,409.875,380.555999755859,400.700012207031,380.955932617188,37581751,0,0
2022-10-10 00:00:00+01:00,395.799987792969,404.470001220703,366.700012207031,394.299987792969,374.871276855469,52952323,0,0
2022-10-17 00:00:00+01:00,394.299987792969,414.799987792969,393,406.5,386.470123291016,26441475,0,0
2022-10-24 00:00:00+01:00,407.100006103516,418.227996826172,407.100006103516,413.299987792969,392.93505859375,26239756,0,0
2022-10-31 00:00:00+00:00,413.899993896484,430.200012207031,412,429.299987792969,408.146667480469,23168047,0,0
2022-11-07 00:00:00+00:00,427.299987792969,445.899993896484,420.652008056641,438.399993896484,416.798278808594,36709117,0,0
2022-11-14 00:00:00+00:00,438.299987792969,458.489990234375,435,455.100006103516,432.675415039063,29106506,0,0
2022-11-21 00:00:00+00:00,454.399993896484,461,450,456.600006103516,434.101501464844,21667730,0,0
2022-11-28 00:00:00+00:00,453.799987792969,456.899993896484,435.100006103516,444.799987792969,422.882934570313,33326204,0,0
2022-12-05 00:00:00+00:00,442.899993896484,450.25,441.299987792969,448,425.925262451172,29147089,0,0
2022-12-12 00:00:00+00:00,445.100006103516,451.299987792969,431.200012207031,436.100006103516,414.611633300781,46593233,0,0
2022-12-19 00:00:00+00:00,436,452.600006103516,433.600006103516,444,422.122344970703,20982140,0,0
2022-12-26 00:00:00+00:00,444,452.058013916016,442.399993896484,442.799987792969,420.981475830078,8249664,0,0
2023-01-02 00:00:00+00:00,445.899993896484,458.149993896484,443.299987792969,456,433.531066894531,28687622,0,0
2023-01-09 00:00:00+00:00,456,461.066009521484,435.799987792969,444.200012207031,422.3125,39237336,0,0
2023-01-16 00:00:00+00:00,444.299987792969,447.200012207031,434.399993896484,439,417.368713378906,35267336,0,0
2023-01-23 00:00:00+00:00,440,459.299987792969,439.5,457.399993896484,434.862091064453,37495012,0,0
2023-01-30 00:00:00+00:00,454.399993896484,459.399993896484,447.799987792969,450.299987792969,428.111907958984,48879358,0,0
2023-02-06 00:00:00+00:00,448,449.200012207031,436.299987792969,440,418.319458007813,38799772,0,0
2023-02-13 00:00:00+00:00,441.200012207031,450.299987792969,440,447.600006103516,425.544982910156,30251441,0,0
2023-02-20 00:00:00+00:00,448.5,450.799987792969,434.299987792969,440,418.319458007813,26764528,0,0
2023-02-27 00:00:00+00:00,442.899993896484,450.5,441.608001708984,447.200012207031,425.164703369141,29895454,0,0
2023-03-06 00:00:00+00:00,447.399993896484,467.299987792969,443.100006103516,449.700012207031,427.54150390625,82322819,0,0
2023-03-13 00:00:00+00:00,450,451.417999267578,400.68701171875,402.200012207031,382.382019042969,85158023,0,0
2023-03-20 00:00:00+00:00,396.200012207031,425.399993896484,383.496002197266,408.299987792969,388.181427001953,60152666,0,0
2023-03-27 00:00:00+01:00,416,422.049987792969,399.549987792969,404.200012207031,384.283477783203,81534829,20.7,0
2023-04-03 00:00:00+01:00,405,434.100006103516,404.399993896484,417.100006103516,417.100006103516,43217151,0,0
2023-04-10 00:00:00+01:00,419.100006103516,426.700012207031,419.100006103516,421.700012207031,421.700012207031,32435695,0,0
2023-04-17 00:00:00+01:00,423.700012207031,427.635009765625,415.399993896484,420.299987792969,420.299987792969,37715986,0,0
2023-04-24 00:00:00+01:00,418.100006103516,423,415.299987792969,423,423,34331974,0,0
2023-05-01 00:00:00+01:00,423.399993896484,426.100006103516,406.399993896484,414.600006103516,414.600006103516,40446519,0,0
2023-05-08 00:00:00+01:00,414.600006103516,419.100006103516,408,412.700012207031,412.700012207031,36950836,0,0
2023-05-15 00:00:00+01:00,414,418.399993896484,407.399993896484,413.5,413.5,53109487,0,0
2023-05-22 00:00:00+01:00,413.600006103516,424,394.700012207031,401.299987792969,401.299987792969,64363368,0,0
2023-05-29 00:00:00+01:00,401.299987792969,409.477996826172,392.700012207031,409.100006103516,409.100006103516,47587959,0,0
2023-06-05 00:00:00+01:00,406.299987792969,410.700012207031,400.100006103516,400.899993896484,400.899993896484,22494985,0,0
2023-06-12 00:00:00+01:00,404.100006103516,406,394.5,396,396,41531163,0,0
2023-06-19 00:00:00+01:00,394,399.899993896484,380.720001220703,386.200012207031,386.200012207031,40439880,0,0
2023-06-26 00:00:00+01:00,387.200012207031,397,382.899993896484,395.200012207031,395.200012207031,27701915,0,0
2023-07-03 00:00:00+01:00,396.5,399.799987792969,380.100006103516,381.799987792969,381.799987792969,26005305,0,0
2023-07-10 00:00:00+01:00,380,392.299987792969,379.403991699219,386,386,29789300,0,0
2023-07-17 00:00:00+01:00,385,389.5,384.251007080078,387.100006103516,387.100006103516,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2021-12-13 00:00:00+00:00 393.999975585938 406.6 391.4 402.899916992188 291.232287597656 62714764.4736842 0 0
3 2021-12-20 00:00:00+00:00 393.999975585938 412.199990234375 392.502983398438 409.899997558594 296.292243652344 46596651.3157895 0 0
4 2021-12-27 00:00:00+00:00 409.899997558594 416.550971679688 408.387001953125 410.4 296.653642578125 10818482.8947368 0 0
5 2022-01-03 00:00:00+00:00 410.4 432.199995117188 410.4 432.099985351563 312.339265136719 44427327.6315789 0 0
6 2022-01-10 00:00:00+00:00 431.3 439.199982910156 429.099970703125 436.099912109375 315.230618896484 29091400 0 0
7 2022-01-17 00:00:00+00:00 437.999912109375 445.199965820313 426.999997558594 431.999975585938 312.267017822266 43787351.3157895 0 0
8 2022-01-24 00:00:00+00:00 430.099975585938 440.999973144531 420.999968261719 433.499982910156 313.351237792969 58487296.0526316 0 0
9 2022-01-31 00:00:00+00:00 436.199968261719 443.049987792969 432.099985351563 435.199916992188 314.580045166016 43335806.5789474 0 0
10 2022-02-07 00:00:00+00:00 437.899995117188 448.799992675781 436.051994628906 444.39998046875 321.230207519531 39644061.8421053 0 0
11 2022-02-14 00:00:00+00:00 437.699975585938 441.999978027344 426.699968261719 432.199995117188 312.411558837891 49972693.4210526 0 0
12 2022-02-21 00:00:00+00:00 435.499992675781 438.476999511719 408.29998046875 423.399970703125 306.050571289063 65719596.0526316 0 0
13 2022-02-28 00:00:00+00:00 415.099995117188 427.999909667969 386.199932861328 386.799945068359 279.594578857422 94057936.8421053 4.1875 0
14 2022-03-07 00:00:00+00:00 374.999952392578 417.299978027344 361.101981201172 409.599968261719 298.389248046875 71269101.3157895 0 0
15 2022-03-14 00:00:00+00:00 413.099985351563 426.699968261719 408.899992675781 422.399965820313 307.713929443359 55431927.6315789 0 0
16 2022-03-21 00:00:00+00:00 422.699995117188 442.7 422.399965820313 437.799985351563 318.932696533203 39896352.6315789 0 0
17 2022-03-28 00:00:00+01:00 442.49998046875 460.999978027344 440.097983398438 444.6 323.886403808594 56413515.7894737 0 0
18 2022-04-04 00:00:00+01:00 439.699985351563 445.399985351563 421.999973144531 425.799973144531 310.190817871094 49415836.8421053 19.342106 0
19 2022-04-11 00:00:00+01:00 425.39998046875 435.599909667969 420.799995117188 434.299968261719 327.211427001953 29875081.5789474 0 0
20 2022-04-18 00:00:00+01:00 434.299968261719 447.799987792969 433.599992675781 437.799985351563 329.848419189453 49288272.3684211 0 0
21 2022-04-25 00:00:00+01:00 430.699987792969 438.799990234375 423.999982910156 433.299916992188 326.457967529297 44656776.3157895 0 0
22 2022-05-02 00:00:00+01:00 433.299916992188 450.999975585938 414.499982910156 414.899975585938 312.595018310547 29538167.1052632 0 0
23 2022-05-09 00:00:00+01:00 413.199995117188 417.449992675781 368.282923583984 408.199970703125 307.547099609375 73989611.8421053 0 0
24 2022-05-16 00:00:00+01:00 384 423.600006103516 384 412.100006103516 310.485473632813 81938261 101.69 0.76
25 2022-05-23 00:00:00+01:00 416.100006103516 442.399993896484 341.915008544922 440.899993896484 409.764678955078 45432941 0 0
26 2022-05-30 00:00:00+01:00 442.700012207031 444.200012207031 426.600006103516 428.700012207031 398.426239013672 37906659 0 0
27 2022-06-06 00:00:00+01:00 425.299987792969 434.010009765625 405.200012207031 405.399993896484 376.771606445313 40648810 0 0
28 2022-06-13 00:00:00+01:00 402.5 420 399.799987792969 411.200012207031 382.162048339844 74196958 0 0
29 2022-06-20 00:00:00+01:00 412.5 421.899993896484 398.399993896484 411.5 382.440826416016 28679717 0 0
30 2022-06-27 00:00:00+01:00 413.100006103516 422.399993896484 397.399993896484 401.600006103516 373.239959716797 35468994 0 0
31 2022-07-04 00:00:00+01:00 405.399993896484 406.600006103516 382.299987792969 401.299987792969 372.961120605469 35304748 0 0
32 2022-07-11 00:00:00+01:00 394.799987792969 405.850006103516 383.399993896484 396.600006103516 368.593048095703 42308459 0 0
33 2022-07-18 00:00:00+01:00 392.5 399.700012207031 384.799987792969 391.700012207031 364.039093017578 36656839 0 0
34 2022-07-25 00:00:00+01:00 392.200012207031 400.799987792969 388.700012207031 396 368.035430908203 33124660 0 0
35 2022-08-01 00:00:00+01:00 396.399993896484 405.5 390.415008544922 402 373.611724853516 21753121 0 0
36 2022-08-08 00:00:00+01:00 406.600006103516 473.700012207031 403.299987792969 467.899993896484 434.858032226563 59155709 0 0
37 2022-08-15 00:00:00+01:00 468.100006103516 470.5 434 437 406.140106201172 36989620 10.3 0
38 2022-08-22 00:00:00+01:00 436.100006103516 436.869995117188 419.299987792969 420.5 399.780303955078 36492572 0 0
39 2022-08-29 00:00:00+01:00 420.5 426.600006103516 408.600006103516 426.600006103516 405.579742431641 29573657 0 0
40 2022-09-05 00:00:00+01:00 418.5 444.4169921875 416.100006103516 443.100006103516 421.266723632813 34375126 0 0
41 2022-09-12 00:00:00+01:00 444.649993896484 448.899993896484 435.200012207031 440.100006103516 418.414520263672 39085960 0 0
42 2022-09-19 00:00:00+01:00 440.100006103516 447.200012207031 419.299987792969 422.899993896484 402.062042236328 27982081 0 0
43 2022-09-26 00:00:00+01:00 421.200012207031 421.200012207031 373.31201171875 388.200012207031 369.071868896484 70408935 0 0
44 2022-10-03 00:00:00+01:00 382.899993896484 409.875 380.555999755859 400.700012207031 380.955932617188 37581751 0 0
45 2022-10-10 00:00:00+01:00 395.799987792969 404.470001220703 366.700012207031 394.299987792969 374.871276855469 52952323 0 0
46 2022-10-17 00:00:00+01:00 394.299987792969 414.799987792969 393 406.5 386.470123291016 26441475 0 0
47 2022-10-24 00:00:00+01:00 407.100006103516 418.227996826172 407.100006103516 413.299987792969 392.93505859375 26239756 0 0
48 2022-10-31 00:00:00+00:00 413.899993896484 430.200012207031 412 429.299987792969 408.146667480469 23168047 0 0
49 2022-11-07 00:00:00+00:00 427.299987792969 445.899993896484 420.652008056641 438.399993896484 416.798278808594 36709117 0 0
50 2022-11-14 00:00:00+00:00 438.299987792969 458.489990234375 435 455.100006103516 432.675415039063 29106506 0 0
51 2022-11-21 00:00:00+00:00 454.399993896484 461 450 456.600006103516 434.101501464844 21667730 0 0
52 2022-11-28 00:00:00+00:00 453.799987792969 456.899993896484 435.100006103516 444.799987792969 422.882934570313 33326204 0 0
53 2022-12-05 00:00:00+00:00 442.899993896484 450.25 441.299987792969 448 425.925262451172 29147089 0 0
54 2022-12-12 00:00:00+00:00 445.100006103516 451.299987792969 431.200012207031 436.100006103516 414.611633300781 46593233 0 0
55 2022-12-19 00:00:00+00:00 436 452.600006103516 433.600006103516 444 422.122344970703 20982140 0 0
56 2022-12-26 00:00:00+00:00 444 452.058013916016 442.399993896484 442.799987792969 420.981475830078 8249664 0 0
57 2023-01-02 00:00:00+00:00 445.899993896484 458.149993896484 443.299987792969 456 433.531066894531 28687622 0 0
58 2023-01-09 00:00:00+00:00 456 461.066009521484 435.799987792969 444.200012207031 422.3125 39237336 0 0
59 2023-01-16 00:00:00+00:00 444.299987792969 447.200012207031 434.399993896484 439 417.368713378906 35267336 0 0
60 2023-01-23 00:00:00+00:00 440 459.299987792969 439.5 457.399993896484 434.862091064453 37495012 0 0
61 2023-01-30 00:00:00+00:00 454.399993896484 459.399993896484 447.799987792969 450.299987792969 428.111907958984 48879358 0 0
62 2023-02-06 00:00:00+00:00 448 449.200012207031 436.299987792969 440 418.319458007813 38799772 0 0
63 2023-02-13 00:00:00+00:00 441.200012207031 450.299987792969 440 447.600006103516 425.544982910156 30251441 0 0
64 2023-02-20 00:00:00+00:00 448.5 450.799987792969 434.299987792969 440 418.319458007813 26764528 0 0
65 2023-02-27 00:00:00+00:00 442.899993896484 450.5 441.608001708984 447.200012207031 425.164703369141 29895454 0 0
66 2023-03-06 00:00:00+00:00 447.399993896484 467.299987792969 443.100006103516 449.700012207031 427.54150390625 82322819 0 0
67 2023-03-13 00:00:00+00:00 450 451.417999267578 400.68701171875 402.200012207031 382.382019042969 85158023 0 0
68 2023-03-20 00:00:00+00:00 396.200012207031 425.399993896484 383.496002197266 408.299987792969 388.181427001953 60152666 0 0
69 2023-03-27 00:00:00+01:00 416 422.049987792969 399.549987792969 404.200012207031 384.283477783203 81534829 20.7 0
70 2023-04-03 00:00:00+01:00 405 434.100006103516 404.399993896484 417.100006103516 417.100006103516 43217151 0 0
71 2023-04-10 00:00:00+01:00 419.100006103516 426.700012207031 419.100006103516 421.700012207031 421.700012207031 32435695 0 0
72 2023-04-17 00:00:00+01:00 423.700012207031 427.635009765625 415.399993896484 420.299987792969 420.299987792969 37715986 0 0
73 2023-04-24 00:00:00+01:00 418.100006103516 423 415.299987792969 423 423 34331974 0 0
74 2023-05-01 00:00:00+01:00 423.399993896484 426.100006103516 406.399993896484 414.600006103516 414.600006103516 40446519 0 0
75 2023-05-08 00:00:00+01:00 414.600006103516 419.100006103516 408 412.700012207031 412.700012207031 36950836 0 0
76 2023-05-15 00:00:00+01:00 414 418.399993896484 407.399993896484 413.5 413.5 53109487 0 0
77 2023-05-22 00:00:00+01:00 413.600006103516 424 394.700012207031 401.299987792969 401.299987792969 64363368 0 0
78 2023-05-29 00:00:00+01:00 401.299987792969 409.477996826172 392.700012207031 409.100006103516 409.100006103516 47587959 0 0
79 2023-06-05 00:00:00+01:00 406.299987792969 410.700012207031 400.100006103516 400.899993896484 400.899993896484 22494985 0 0
80 2023-06-12 00:00:00+01:00 404.100006103516 406 394.5 396 396 41531163 0 0
81 2023-06-19 00:00:00+01:00 394 399.899993896484 380.720001220703 386.200012207031 386.200012207031 40439880 0 0
82 2023-06-26 00:00:00+01:00 387.200012207031 397 382.899993896484 395.200012207031 395.200012207031 27701915 0 0
83 2023-07-03 00:00:00+01:00 396.5 399.799987792969 380.100006103516 381.799987792969 381.799987792969 26005305 0 0
84 2023-07-10 00:00:00+01:00 380 392.299987792969 379.403991699219 386 386 29789300 0 0
85 2023-07-17 00:00:00+01:00 385 389.5 384.251007080078 387.100006103516 387.100006103516 0 0 0

View File

@@ -0,0 +1,85 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2021-12-13 00:00:00+00:00,518.4210205078125,535.0,515.0,530.1314697265625,383.20037841796875,47663221,0.0,0.0
2021-12-20 00:00:00+00:00,518.4210205078125,542.368408203125,516.4512939453125,539.3421020507812,389.85821533203125,35413455,0.0,0.0
2021-12-27 00:00:00+00:00,539.3421020507812,548.0933837890625,537.351318359375,540.0,390.333740234375,8222047,0.0,0.0
2022-01-03 00:00:00+00:00,540.0,568.6842041015625,540.0,568.5526123046875,410.97271728515625,33764769,0.0,0.0
2022-01-10 00:00:00+00:00,567.5,577.8947143554688,564.605224609375,573.815673828125,414.7771301269531,22109464,0.0,0.0
2022-01-17 00:00:00+00:00,576.315673828125,585.7894287109375,561.8421020507812,568.4210205078125,410.8776550292969,33278387,0.0,0.0
2022-01-24 00:00:00+00:00,565.9210205078125,580.2631225585938,553.9473266601562,570.3947143554688,412.30426025390625,44450345,0.0,0.0
2022-01-31 00:00:00+00:00,573.9473266601562,582.9605102539062,568.5526123046875,572.6314697265625,413.9211120605469,32935213,0.0,0.0
2022-02-07 00:00:00+00:00,576.1842041015625,590.5263061523438,573.7526245117188,584.73681640625,422.67132568359375,30129487,0.0,0.0
2022-02-14 00:00:00+00:00,575.9210205078125,581.5789184570312,561.4473266601562,568.6842041015625,411.0678405761719,37979247,0.0,0.0
2022-02-21 00:00:00+00:00,573.0263061523438,576.9434204101562,537.23681640625,557.105224609375,402.6981201171875,49946893,0.0,0.0
2022-02-28 00:00:00+00:00,546.1842041015625,563.1577758789062,508.1578063964844,508.9472961425781,367.8876037597656,71484032,4.1875,0.0
2022-03-07 00:00:00+00:00,493.4209899902344,549.0789184570312,475.1341857910156,538.9473266601562,392.617431640625,54164517,0.0,0.0
2022-03-14 00:00:00+00:00,543.5526123046875,561.4473266601562,538.0263061523438,555.7894287109375,404.8867492675781,42128265,0.0,0.0
2022-03-21 00:00:00+00:00,556.1842041015625,582.5,555.7894287109375,576.0526123046875,419.6482849121094,30321228,0.0,0.0
2022-03-28 00:00:00+01:00,582.23681640625,606.5789184570312,579.0762939453125,585.0,426.16632080078125,42874272,0.0,0.0
2022-04-04 00:00:00+01:00,578.5526123046875,586.0526123046875,555.2631225585938,560.2631225585938,408.14581298828125,37556036,19.342106,0.0
2022-04-11 00:00:00+01:00,559.73681640625,573.1577758789062,553.6842041015625,571.4473266601562,430.5413513183594,22705062,0.0,0.0
2022-04-18 00:00:00+01:00,571.4473266601562,589.2105102539062,570.5263061523438,576.0526123046875,434.0110778808594,37459087,0.0,0.0
2022-04-25 00:00:00+01:00,566.7105102539062,577.368408203125,557.8947143554688,570.1314697265625,429.5499572753906,33939150,0.0,0.0
2022-05-02 00:00:00+01:00,570.1314697265625,593.4210205078125,545.3947143554688,545.9210205078125,411.3092346191406,22449007,0.0,0.0
2022-05-09 00:00:00+01:00,543.6842041015625,549.2763061523438,484.5827941894531,537.105224609375,404.667236328125,56232105,0.0,0.0
2022-05-16 00:00:00+01:00,384.0,423.6000061035156,384.0,412.1000061035156,310.4854736328125,81938261,101.69,0.76
2022-05-23 00:00:00+01:00,416.1000061035156,442.3999938964844,341.9150085449219,440.8999938964844,409.7646789550781,45432941,0.0,0.0
2022-05-30 00:00:00+01:00,442.70001220703125,444.20001220703125,426.6000061035156,428.70001220703125,398.4262390136719,37906659,0.0,0.0
2022-06-06 00:00:00+01:00,425.29998779296875,434.010009765625,405.20001220703125,405.3999938964844,376.7716064453125,40648810,0.0,0.0
2022-06-13 00:00:00+01:00,402.5,420.0,399.79998779296875,411.20001220703125,382.16204833984375,74196958,0.0,0.0
2022-06-20 00:00:00+01:00,412.5,421.8999938964844,398.3999938964844,411.5,382.4408264160156,28679717,0.0,0.0
2022-06-27 00:00:00+01:00,413.1000061035156,422.3999938964844,397.3999938964844,401.6000061035156,373.2399597167969,35468994,0.0,0.0
2022-07-04 00:00:00+01:00,405.3999938964844,406.6000061035156,382.29998779296875,401.29998779296875,372.96112060546875,35304748,0.0,0.0
2022-07-11 00:00:00+01:00,394.79998779296875,405.8500061035156,383.3999938964844,396.6000061035156,368.5930480957031,42308459,0.0,0.0
2022-07-18 00:00:00+01:00,392.5,399.70001220703125,384.79998779296875,391.70001220703125,364.0390930175781,36656839,0.0,0.0
2022-07-25 00:00:00+01:00,392.20001220703125,400.79998779296875,388.70001220703125,396.0,368.0354309082031,33124660,0.0,0.0
2022-08-01 00:00:00+01:00,396.3999938964844,405.5,390.4150085449219,402.0,373.6117248535156,21753121,0.0,0.0
2022-08-08 00:00:00+01:00,406.6000061035156,473.70001220703125,403.29998779296875,467.8999938964844,434.8580322265625,59155709,0.0,0.0
2022-08-15 00:00:00+01:00,468.1000061035156,470.5,434.0,437.0,406.1401062011719,36989620,10.3,0.0
2022-08-22 00:00:00+01:00,436.1000061035156,436.8699951171875,419.29998779296875,420.5,399.7803039550781,36492572,0.0,0.0
2022-08-29 00:00:00+01:00,420.5,426.6000061035156,408.6000061035156,426.6000061035156,405.5797424316406,29573657,0.0,0.0
2022-09-05 00:00:00+01:00,418.5,444.4169921875,416.1000061035156,443.1000061035156,421.2667236328125,34375126,0.0,0.0
2022-09-12 00:00:00+01:00,444.6499938964844,448.8999938964844,435.20001220703125,440.1000061035156,418.4145202636719,39085960,0.0,0.0
2022-09-19 00:00:00+01:00,440.1000061035156,447.20001220703125,419.29998779296875,422.8999938964844,402.0620422363281,27982081,0.0,0.0
2022-09-26 00:00:00+01:00,421.20001220703125,421.20001220703125,373.31201171875,388.20001220703125,369.0718688964844,70408935,0.0,0.0
2022-10-03 00:00:00+01:00,382.8999938964844,409.875,380.5559997558594,400.70001220703125,380.9559326171875,37581751,0.0,0.0
2022-10-10 00:00:00+01:00,395.79998779296875,404.4700012207031,366.70001220703125,394.29998779296875,374.87127685546875,52952323,0.0,0.0
2022-10-17 00:00:00+01:00,394.29998779296875,414.79998779296875,393.0,406.5,386.4701232910156,26441475,0.0,0.0
2022-10-24 00:00:00+01:00,407.1000061035156,418.2279968261719,407.1000061035156,413.29998779296875,392.93505859375,26239756,0.0,0.0
2022-10-31 00:00:00+00:00,413.8999938964844,430.20001220703125,412.0,429.29998779296875,408.14666748046875,23168047,0.0,0.0
2022-11-07 00:00:00+00:00,427.29998779296875,445.8999938964844,420.6520080566406,438.3999938964844,416.79827880859375,36709117,0.0,0.0
2022-11-14 00:00:00+00:00,438.29998779296875,458.489990234375,435.0,455.1000061035156,432.6754150390625,29106506,0.0,0.0
2022-11-21 00:00:00+00:00,454.3999938964844,461.0,450.0,456.6000061035156,434.10150146484375,21667730,0.0,0.0
2022-11-28 00:00:00+00:00,453.79998779296875,456.8999938964844,435.1000061035156,444.79998779296875,422.8829345703125,33326204,0.0,0.0
2022-12-05 00:00:00+00:00,442.8999938964844,450.25,441.29998779296875,448.0,425.9252624511719,29147089,0.0,0.0
2022-12-12 00:00:00+00:00,445.1000061035156,451.29998779296875,431.20001220703125,436.1000061035156,414.61163330078125,46593233,0.0,0.0
2022-12-19 00:00:00+00:00,436.0,452.6000061035156,433.6000061035156,444.0,422.1223449707031,20982140,0.0,0.0
2022-12-26 00:00:00+00:00,444.0,452.0580139160156,442.3999938964844,442.79998779296875,420.9814758300781,8249664,0.0,0.0
2023-01-02 00:00:00+00:00,445.8999938964844,458.1499938964844,443.29998779296875,456.0,433.53106689453125,28687622,0.0,0.0
2023-01-09 00:00:00+00:00,456.0,461.0660095214844,435.79998779296875,444.20001220703125,422.3125,39237336,0.0,0.0
2023-01-16 00:00:00+00:00,444.29998779296875,447.20001220703125,434.3999938964844,439.0,417.36871337890625,35267336,0.0,0.0
2023-01-23 00:00:00+00:00,440.0,459.29998779296875,439.5,457.3999938964844,434.8620910644531,37495012,0.0,0.0
2023-01-30 00:00:00+00:00,454.3999938964844,459.3999938964844,447.79998779296875,450.29998779296875,428.1119079589844,48879358,0.0,0.0
2023-02-06 00:00:00+00:00,448.0,449.20001220703125,436.29998779296875,440.0,418.3194580078125,38799772,0.0,0.0
2023-02-13 00:00:00+00:00,441.20001220703125,450.29998779296875,440.0,447.6000061035156,425.54498291015625,30251441,0.0,0.0
2023-02-20 00:00:00+00:00,448.5,450.79998779296875,434.29998779296875,440.0,418.3194580078125,26764528,0.0,0.0
2023-02-27 00:00:00+00:00,442.8999938964844,450.5,441.6080017089844,447.20001220703125,425.1647033691406,29895454,0.0,0.0
2023-03-06 00:00:00+00:00,447.3999938964844,467.29998779296875,443.1000061035156,449.70001220703125,427.54150390625,82322819,0.0,0.0
2023-03-13 00:00:00+00:00,450.0,451.4179992675781,400.68701171875,402.20001220703125,382.38201904296875,85158023,0.0,0.0
2023-03-20 00:00:00+00:00,396.20001220703125,425.3999938964844,383.4960021972656,408.29998779296875,388.1814270019531,60152666,0.0,0.0
2023-03-27 00:00:00+01:00,416.0,422.04998779296875,399.54998779296875,404.20001220703125,384.2834777832031,81534829,20.7,0.0
2023-04-03 00:00:00+01:00,405.0,434.1000061035156,404.3999938964844,417.1000061035156,417.1000061035156,43217151,0.0,0.0
2023-04-10 00:00:00+01:00,419.1000061035156,426.70001220703125,419.1000061035156,421.70001220703125,421.70001220703125,32435695,0.0,0.0
2023-04-17 00:00:00+01:00,423.70001220703125,427.635009765625,415.3999938964844,420.29998779296875,420.29998779296875,37715986,0.0,0.0
2023-04-24 00:00:00+01:00,418.1000061035156,423.0,415.29998779296875,423.0,423.0,34331974,0.0,0.0
2023-05-01 00:00:00+01:00,423.3999938964844,426.1000061035156,406.3999938964844,414.6000061035156,414.6000061035156,40446519,0.0,0.0
2023-05-08 00:00:00+01:00,414.6000061035156,419.1000061035156,408.0,412.70001220703125,412.70001220703125,36950836,0.0,0.0
2023-05-15 00:00:00+01:00,414.0,418.3999938964844,407.3999938964844,413.5,413.5,53109487,0.0,0.0
2023-05-22 00:00:00+01:00,413.6000061035156,424.0,394.70001220703125,401.29998779296875,401.29998779296875,64363368,0.0,0.0
2023-05-29 00:00:00+01:00,401.29998779296875,409.4779968261719,392.70001220703125,409.1000061035156,409.1000061035156,47587959,0.0,0.0
2023-06-05 00:00:00+01:00,406.29998779296875,410.70001220703125,400.1000061035156,400.8999938964844,400.8999938964844,22494985,0.0,0.0
2023-06-12 00:00:00+01:00,404.1000061035156,406.0,394.5,396.0,396.0,41531163,0.0,0.0
2023-06-19 00:00:00+01:00,394.0,399.8999938964844,380.7200012207031,386.20001220703125,386.20001220703125,40439880,0.0,0.0
2023-06-26 00:00:00+01:00,387.20001220703125,397.0,382.8999938964844,395.20001220703125,395.20001220703125,27701915,0.0,0.0
2023-07-03 00:00:00+01:00,396.5,399.79998779296875,380.1000061035156,381.79998779296875,381.79998779296875,26005305,0.0,0.0
2023-07-10 00:00:00+01:00,380.0,392.29998779296875,379.40399169921875,386.0,386.0,29789300,0.0,0.0
2023-07-17 00:00:00+01:00,385.0,389.5,384.2510070800781,387.1000061035156,387.1000061035156,0,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2021-12-13 00:00:00+00:00 518.4210205078125 535.0 515.0 530.1314697265625 383.20037841796875 47663221 0.0 0.0
3 2021-12-20 00:00:00+00:00 518.4210205078125 542.368408203125 516.4512939453125 539.3421020507812 389.85821533203125 35413455 0.0 0.0
4 2021-12-27 00:00:00+00:00 539.3421020507812 548.0933837890625 537.351318359375 540.0 390.333740234375 8222047 0.0 0.0
5 2022-01-03 00:00:00+00:00 540.0 568.6842041015625 540.0 568.5526123046875 410.97271728515625 33764769 0.0 0.0
6 2022-01-10 00:00:00+00:00 567.5 577.8947143554688 564.605224609375 573.815673828125 414.7771301269531 22109464 0.0 0.0
7 2022-01-17 00:00:00+00:00 576.315673828125 585.7894287109375 561.8421020507812 568.4210205078125 410.8776550292969 33278387 0.0 0.0
8 2022-01-24 00:00:00+00:00 565.9210205078125 580.2631225585938 553.9473266601562 570.3947143554688 412.30426025390625 44450345 0.0 0.0
9 2022-01-31 00:00:00+00:00 573.9473266601562 582.9605102539062 568.5526123046875 572.6314697265625 413.9211120605469 32935213 0.0 0.0
10 2022-02-07 00:00:00+00:00 576.1842041015625 590.5263061523438 573.7526245117188 584.73681640625 422.67132568359375 30129487 0.0 0.0
11 2022-02-14 00:00:00+00:00 575.9210205078125 581.5789184570312 561.4473266601562 568.6842041015625 411.0678405761719 37979247 0.0 0.0
12 2022-02-21 00:00:00+00:00 573.0263061523438 576.9434204101562 537.23681640625 557.105224609375 402.6981201171875 49946893 0.0 0.0
13 2022-02-28 00:00:00+00:00 546.1842041015625 563.1577758789062 508.1578063964844 508.9472961425781 367.8876037597656 71484032 4.1875 0.0
14 2022-03-07 00:00:00+00:00 493.4209899902344 549.0789184570312 475.1341857910156 538.9473266601562 392.617431640625 54164517 0.0 0.0
15 2022-03-14 00:00:00+00:00 543.5526123046875 561.4473266601562 538.0263061523438 555.7894287109375 404.8867492675781 42128265 0.0 0.0
16 2022-03-21 00:00:00+00:00 556.1842041015625 582.5 555.7894287109375 576.0526123046875 419.6482849121094 30321228 0.0 0.0
17 2022-03-28 00:00:00+01:00 582.23681640625 606.5789184570312 579.0762939453125 585.0 426.16632080078125 42874272 0.0 0.0
18 2022-04-04 00:00:00+01:00 578.5526123046875 586.0526123046875 555.2631225585938 560.2631225585938 408.14581298828125 37556036 19.342106 0.0
19 2022-04-11 00:00:00+01:00 559.73681640625 573.1577758789062 553.6842041015625 571.4473266601562 430.5413513183594 22705062 0.0 0.0
20 2022-04-18 00:00:00+01:00 571.4473266601562 589.2105102539062 570.5263061523438 576.0526123046875 434.0110778808594 37459087 0.0 0.0
21 2022-04-25 00:00:00+01:00 566.7105102539062 577.368408203125 557.8947143554688 570.1314697265625 429.5499572753906 33939150 0.0 0.0
22 2022-05-02 00:00:00+01:00 570.1314697265625 593.4210205078125 545.3947143554688 545.9210205078125 411.3092346191406 22449007 0.0 0.0
23 2022-05-09 00:00:00+01:00 543.6842041015625 549.2763061523438 484.5827941894531 537.105224609375 404.667236328125 56232105 0.0 0.0
24 2022-05-16 00:00:00+01:00 384.0 423.6000061035156 384.0 412.1000061035156 310.4854736328125 81938261 101.69 0.76
25 2022-05-23 00:00:00+01:00 416.1000061035156 442.3999938964844 341.9150085449219 440.8999938964844 409.7646789550781 45432941 0.0 0.0
26 2022-05-30 00:00:00+01:00 442.70001220703125 444.20001220703125 426.6000061035156 428.70001220703125 398.4262390136719 37906659 0.0 0.0
27 2022-06-06 00:00:00+01:00 425.29998779296875 434.010009765625 405.20001220703125 405.3999938964844 376.7716064453125 40648810 0.0 0.0
28 2022-06-13 00:00:00+01:00 402.5 420.0 399.79998779296875 411.20001220703125 382.16204833984375 74196958 0.0 0.0
29 2022-06-20 00:00:00+01:00 412.5 421.8999938964844 398.3999938964844 411.5 382.4408264160156 28679717 0.0 0.0
30 2022-06-27 00:00:00+01:00 413.1000061035156 422.3999938964844 397.3999938964844 401.6000061035156 373.2399597167969 35468994 0.0 0.0
31 2022-07-04 00:00:00+01:00 405.3999938964844 406.6000061035156 382.29998779296875 401.29998779296875 372.96112060546875 35304748 0.0 0.0
32 2022-07-11 00:00:00+01:00 394.79998779296875 405.8500061035156 383.3999938964844 396.6000061035156 368.5930480957031 42308459 0.0 0.0
33 2022-07-18 00:00:00+01:00 392.5 399.70001220703125 384.79998779296875 391.70001220703125 364.0390930175781 36656839 0.0 0.0
34 2022-07-25 00:00:00+01:00 392.20001220703125 400.79998779296875 388.70001220703125 396.0 368.0354309082031 33124660 0.0 0.0
35 2022-08-01 00:00:00+01:00 396.3999938964844 405.5 390.4150085449219 402.0 373.6117248535156 21753121 0.0 0.0
36 2022-08-08 00:00:00+01:00 406.6000061035156 473.70001220703125 403.29998779296875 467.8999938964844 434.8580322265625 59155709 0.0 0.0
37 2022-08-15 00:00:00+01:00 468.1000061035156 470.5 434.0 437.0 406.1401062011719 36989620 10.3 0.0
38 2022-08-22 00:00:00+01:00 436.1000061035156 436.8699951171875 419.29998779296875 420.5 399.7803039550781 36492572 0.0 0.0
39 2022-08-29 00:00:00+01:00 420.5 426.6000061035156 408.6000061035156 426.6000061035156 405.5797424316406 29573657 0.0 0.0
40 2022-09-05 00:00:00+01:00 418.5 444.4169921875 416.1000061035156 443.1000061035156 421.2667236328125 34375126 0.0 0.0
41 2022-09-12 00:00:00+01:00 444.6499938964844 448.8999938964844 435.20001220703125 440.1000061035156 418.4145202636719 39085960 0.0 0.0
42 2022-09-19 00:00:00+01:00 440.1000061035156 447.20001220703125 419.29998779296875 422.8999938964844 402.0620422363281 27982081 0.0 0.0
43 2022-09-26 00:00:00+01:00 421.20001220703125 421.20001220703125 373.31201171875 388.20001220703125 369.0718688964844 70408935 0.0 0.0
44 2022-10-03 00:00:00+01:00 382.8999938964844 409.875 380.5559997558594 400.70001220703125 380.9559326171875 37581751 0.0 0.0
45 2022-10-10 00:00:00+01:00 395.79998779296875 404.4700012207031 366.70001220703125 394.29998779296875 374.87127685546875 52952323 0.0 0.0
46 2022-10-17 00:00:00+01:00 394.29998779296875 414.79998779296875 393.0 406.5 386.4701232910156 26441475 0.0 0.0
47 2022-10-24 00:00:00+01:00 407.1000061035156 418.2279968261719 407.1000061035156 413.29998779296875 392.93505859375 26239756 0.0 0.0
48 2022-10-31 00:00:00+00:00 413.8999938964844 430.20001220703125 412.0 429.29998779296875 408.14666748046875 23168047 0.0 0.0
49 2022-11-07 00:00:00+00:00 427.29998779296875 445.8999938964844 420.6520080566406 438.3999938964844 416.79827880859375 36709117 0.0 0.0
50 2022-11-14 00:00:00+00:00 438.29998779296875 458.489990234375 435.0 455.1000061035156 432.6754150390625 29106506 0.0 0.0
51 2022-11-21 00:00:00+00:00 454.3999938964844 461.0 450.0 456.6000061035156 434.10150146484375 21667730 0.0 0.0
52 2022-11-28 00:00:00+00:00 453.79998779296875 456.8999938964844 435.1000061035156 444.79998779296875 422.8829345703125 33326204 0.0 0.0
53 2022-12-05 00:00:00+00:00 442.8999938964844 450.25 441.29998779296875 448.0 425.9252624511719 29147089 0.0 0.0
54 2022-12-12 00:00:00+00:00 445.1000061035156 451.29998779296875 431.20001220703125 436.1000061035156 414.61163330078125 46593233 0.0 0.0
55 2022-12-19 00:00:00+00:00 436.0 452.6000061035156 433.6000061035156 444.0 422.1223449707031 20982140 0.0 0.0
56 2022-12-26 00:00:00+00:00 444.0 452.0580139160156 442.3999938964844 442.79998779296875 420.9814758300781 8249664 0.0 0.0
57 2023-01-02 00:00:00+00:00 445.8999938964844 458.1499938964844 443.29998779296875 456.0 433.53106689453125 28687622 0.0 0.0
58 2023-01-09 00:00:00+00:00 456.0 461.0660095214844 435.79998779296875 444.20001220703125 422.3125 39237336 0.0 0.0
59 2023-01-16 00:00:00+00:00 444.29998779296875 447.20001220703125 434.3999938964844 439.0 417.36871337890625 35267336 0.0 0.0
60 2023-01-23 00:00:00+00:00 440.0 459.29998779296875 439.5 457.3999938964844 434.8620910644531 37495012 0.0 0.0
61 2023-01-30 00:00:00+00:00 454.3999938964844 459.3999938964844 447.79998779296875 450.29998779296875 428.1119079589844 48879358 0.0 0.0
62 2023-02-06 00:00:00+00:00 448.0 449.20001220703125 436.29998779296875 440.0 418.3194580078125 38799772 0.0 0.0
63 2023-02-13 00:00:00+00:00 441.20001220703125 450.29998779296875 440.0 447.6000061035156 425.54498291015625 30251441 0.0 0.0
64 2023-02-20 00:00:00+00:00 448.5 450.79998779296875 434.29998779296875 440.0 418.3194580078125 26764528 0.0 0.0
65 2023-02-27 00:00:00+00:00 442.8999938964844 450.5 441.6080017089844 447.20001220703125 425.1647033691406 29895454 0.0 0.0
66 2023-03-06 00:00:00+00:00 447.3999938964844 467.29998779296875 443.1000061035156 449.70001220703125 427.54150390625 82322819 0.0 0.0
67 2023-03-13 00:00:00+00:00 450.0 451.4179992675781 400.68701171875 402.20001220703125 382.38201904296875 85158023 0.0 0.0
68 2023-03-20 00:00:00+00:00 396.20001220703125 425.3999938964844 383.4960021972656 408.29998779296875 388.1814270019531 60152666 0.0 0.0
69 2023-03-27 00:00:00+01:00 416.0 422.04998779296875 399.54998779296875 404.20001220703125 384.2834777832031 81534829 20.7 0.0
70 2023-04-03 00:00:00+01:00 405.0 434.1000061035156 404.3999938964844 417.1000061035156 417.1000061035156 43217151 0.0 0.0
71 2023-04-10 00:00:00+01:00 419.1000061035156 426.70001220703125 419.1000061035156 421.70001220703125 421.70001220703125 32435695 0.0 0.0
72 2023-04-17 00:00:00+01:00 423.70001220703125 427.635009765625 415.3999938964844 420.29998779296875 420.29998779296875 37715986 0.0 0.0
73 2023-04-24 00:00:00+01:00 418.1000061035156 423.0 415.29998779296875 423.0 423.0 34331974 0.0 0.0
74 2023-05-01 00:00:00+01:00 423.3999938964844 426.1000061035156 406.3999938964844 414.6000061035156 414.6000061035156 40446519 0.0 0.0
75 2023-05-08 00:00:00+01:00 414.6000061035156 419.1000061035156 408.0 412.70001220703125 412.70001220703125 36950836 0.0 0.0
76 2023-05-15 00:00:00+01:00 414.0 418.3999938964844 407.3999938964844 413.5 413.5 53109487 0.0 0.0
77 2023-05-22 00:00:00+01:00 413.6000061035156 424.0 394.70001220703125 401.29998779296875 401.29998779296875 64363368 0.0 0.0
78 2023-05-29 00:00:00+01:00 401.29998779296875 409.4779968261719 392.70001220703125 409.1000061035156 409.1000061035156 47587959 0.0 0.0
79 2023-06-05 00:00:00+01:00 406.29998779296875 410.70001220703125 400.1000061035156 400.8999938964844 400.8999938964844 22494985 0.0 0.0
80 2023-06-12 00:00:00+01:00 404.1000061035156 406.0 394.5 396.0 396.0 41531163 0.0 0.0
81 2023-06-19 00:00:00+01:00 394.0 399.8999938964844 380.7200012207031 386.20001220703125 386.20001220703125 40439880 0.0 0.0
82 2023-06-26 00:00:00+01:00 387.20001220703125 397.0 382.8999938964844 395.20001220703125 395.20001220703125 27701915 0.0 0.0
83 2023-07-03 00:00:00+01:00 396.5 399.79998779296875 380.1000061035156 381.79998779296875 381.79998779296875 26005305 0.0 0.0
84 2023-07-10 00:00:00+01:00 380.0 392.29998779296875 379.40399169921875 386.0 386.0 29789300 0.0 0.0
85 2023-07-17 00:00:00+01:00 385.0 389.5 384.2510070800781 387.1000061035156 387.1000061035156 0 0.0 0.0

View File

@@ -0,0 +1,11 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-18 00:00:00+01:00,193.220001220703,200.839996337891,193.220001220703,196.839996337891,196.839996337891,653125,0,0
2023-05-17 00:00:00+01:00,199.740005493164,207.738006591797,190.121994018555,197.860000610352,197.860000610352,822268,0,0
2023-05-16 00:00:00+01:00,215.600006103516,215.600006103516,201.149993896484,205.100006103516,205.100006103516,451009,243.93939,0.471428571428571
2023-05-15 00:00:00+01:00,456.9090,464.9696,446.7272,461.1515,217.2121,830506.0000,0,0
2023-05-12 00:00:00+01:00,455.2121,458.6060,444.6060,448.4242,211.2173,717655.0000,0,0
2023-05-11 00:00:00+01:00,466.6666,466.6666,450.1212,456.0606,214.8142,1682077.0000,0,0
2023-05-10 00:00:00+01:00,462.8484,473.0303,450.9696,456.9090,215.2138,2639957.0000,0,0
2023-05-09 00:00:00+01:00,475.1515,482.9746,462.8485,463.2727,218.2112,898585.2857,0,0
2023-05-05 00:00:00+01:00,468.7878,477.6969,468.3636,476.0000,224.2061,454704.0000,0,0
2023-05-04 00:00:00+01:00,460.3030,472.6060,460.0527,469.6363,221.2086,415321.0000,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-18 00:00:00+01:00 193.220001220703 200.839996337891 193.220001220703 196.839996337891 196.839996337891 653125 0 0
3 2023-05-17 00:00:00+01:00 199.740005493164 207.738006591797 190.121994018555 197.860000610352 197.860000610352 822268 0 0
4 2023-05-16 00:00:00+01:00 215.600006103516 215.600006103516 201.149993896484 205.100006103516 205.100006103516 451009 243.93939 0.471428571428571
5 2023-05-15 00:00:00+01:00 456.9090 464.9696 446.7272 461.1515 217.2121 830506.0000 0 0
6 2023-05-12 00:00:00+01:00 455.2121 458.6060 444.6060 448.4242 211.2173 717655.0000 0 0
7 2023-05-11 00:00:00+01:00 466.6666 466.6666 450.1212 456.0606 214.8142 1682077.0000 0 0
8 2023-05-10 00:00:00+01:00 462.8484 473.0303 450.9696 456.9090 215.2138 2639957.0000 0 0
9 2023-05-09 00:00:00+01:00 475.1515 482.9746 462.8485 463.2727 218.2112 898585.2857 0 0
10 2023-05-05 00:00:00+01:00 468.7878 477.6969 468.3636 476.0000 224.2061 454704.0000 0 0
11 2023-05-04 00:00:00+01:00 460.3030 472.6060 460.0527 469.6363 221.2086 415321.0000 0 0

View File

@@ -0,0 +1,11 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-18 00:00:00+01:00,193.220001220703,200.839996337891,193.220001220703,196.839996337891,196.839996337891,653125,0,0
2023-05-17 00:00:00+01:00,199.740005493164,207.738006591797,190.121994018555,197.860000610352,197.860000610352,822268,0,0
2023-05-16 00:00:00+01:00,215.600006103516,215.600006103516,201.149993896484,205.100006103516,205.100006103516,451009,243.93939,0.471428571428571
2023-05-15 00:00:00+01:00,456.908996582031,464.969604492188,446.727203369141,461.151489257813,217.21208190918,830506,0,0
2023-05-12 00:00:00+01:00,455.212097167969,458.605987548828,444.605987548828,448.424194335938,211.217269897461,717655,0,0
2023-05-11 00:00:00+01:00,466.666595458984,466.666595458984,450.121185302734,456.060607910156,214.814178466797,1682077,0,0
2023-05-10 00:00:00+01:00,462.848388671875,473.030303955078,450.969604492188,456.908996582031,215.213790893555,2639957,0,0
2023-05-09 00:00:00+01:00,224,227.688003540039,218.199996948242,218.399993896484,102.87100982666,1906090,0,0
2023-05-05 00:00:00+01:00,468.787811279297,477.696899414063,468.363586425781,476,224.2060546875,454704,0,0
2023-05-04 00:00:00+01:00,460.303009033203,472.605987548828,460.052703857422,469.636291503906,221.208602905273,415321,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-18 00:00:00+01:00 193.220001220703 200.839996337891 193.220001220703 196.839996337891 196.839996337891 653125 0 0
3 2023-05-17 00:00:00+01:00 199.740005493164 207.738006591797 190.121994018555 197.860000610352 197.860000610352 822268 0 0
4 2023-05-16 00:00:00+01:00 215.600006103516 215.600006103516 201.149993896484 205.100006103516 205.100006103516 451009 243.93939 0.471428571428571
5 2023-05-15 00:00:00+01:00 456.908996582031 464.969604492188 446.727203369141 461.151489257813 217.21208190918 830506 0 0
6 2023-05-12 00:00:00+01:00 455.212097167969 458.605987548828 444.605987548828 448.424194335938 211.217269897461 717655 0 0
7 2023-05-11 00:00:00+01:00 466.666595458984 466.666595458984 450.121185302734 456.060607910156 214.814178466797 1682077 0 0
8 2023-05-10 00:00:00+01:00 462.848388671875 473.030303955078 450.969604492188 456.908996582031 215.213790893555 2639957 0 0
9 2023-05-09 00:00:00+01:00 224 227.688003540039 218.199996948242 218.399993896484 102.87100982666 1906090 0 0
10 2023-05-05 00:00:00+01:00 468.787811279297 477.696899414063 468.363586425781 476 224.2060546875 454704 0 0
11 2023-05-04 00:00:00+01:00 460.303009033203 472.605987548828 460.052703857422 469.636291503906 221.208602905273 415321 0 0

View File

@@ -0,0 +1,24 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-31 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-30 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0.4406
2023-05-29 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-26 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-25 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-24 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-23 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-22 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-19 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-18 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-17 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-16 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-15 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-12 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-11 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-10 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-09 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-08 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-05 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-04 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-03 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-02 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-01 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-31 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
3 2023-05-30 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0.4406
4 2023-05-29 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
5 2023-05-26 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
6 2023-05-25 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
7 2023-05-24 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
8 2023-05-23 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
9 2023-05-22 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
10 2023-05-19 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
11 2023-05-18 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
12 2023-05-17 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
13 2023-05-16 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
14 2023-05-15 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
15 2023-05-12 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
16 2023-05-11 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
17 2023-05-10 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
18 2023-05-09 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
19 2023-05-08 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
20 2023-05-05 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
21 2023-05-04 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
22 2023-05-03 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
23 2023-05-02 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
24 2023-05-01 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0

View File

@@ -0,0 +1,24 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-31 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-30 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0.4406
2023-05-29 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-26 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-25 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-24 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-23 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-22 00:00:00+10:00,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0.120290003716946,0,0,0
2023-05-19 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-18 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-17 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-16 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-15 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-12 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-11 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-10 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-09 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-08 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-05 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-04 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-03 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-02 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
2023-05-01 00:00:00+10:00,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0.0529999993741512,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-31 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
3 2023-05-30 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0.4406
4 2023-05-29 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
5 2023-05-26 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
6 2023-05-25 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
7 2023-05-24 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
8 2023-05-23 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
9 2023-05-22 00:00:00+10:00 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0.120290003716946 0 0 0
10 2023-05-19 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
11 2023-05-18 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
12 2023-05-17 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
13 2023-05-16 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
14 2023-05-15 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
15 2023-05-12 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
16 2023-05-11 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
17 2023-05-10 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
18 2023-05-09 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
19 2023-05-08 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
20 2023-05-05 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
21 2023-05-04 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
22 2023-05-03 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
23 2023-05-02 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0
24 2023-05-01 00:00:00+10:00 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0.0529999993741512 0 0 0

View File

@@ -0,0 +1,42 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2020-09-30 00:00:00-04:00,4.40000009536743,4.44999980926514,4.01999998092651,4.44999980926514,4.44999980926514,22600,0,0
2020-09-29 00:00:00-04:00,4.3899998664856,4.40000009536743,4.13000011444092,4.30000019073486,4.30000019073486,10800,0,0
2020-09-28 00:00:00-04:00,4.09000015258789,4.25,4.09000015258789,4.25,4.25,8000,0,0
2020-09-25 00:00:00-04:00,3.95000004768372,4.09999990463257,3.95000004768372,4.05000019073486,4.05000019073486,13500,0,0
2020-09-24 00:00:00-04:00,3.84999990463257,4,3.84999990463257,4,4,8800,0,0
2020-09-23 00:00:00-04:00,3.99000000953674,4,3.99000000953674,4,4,5900,0,0
2020-09-22 00:00:00-04:00,3.90000009536743,4.09999990463257,3.84999990463257,4.09999990463257,4.09999990463257,3100,0,0
2020-09-21 00:00:00-04:00,4.09999990463257,4.09999990463257,4.09999990463257,4.09999990463257,4.09999990463257,1200,0,0
2020-09-18 00:00:00-04:00,3.92000007629395,4.09999990463257,3.92000007629395,4.09999990463257,4.09999990463257,27200,0,0
2020-09-17 00:00:00-04:00,3.90000009536743,3.99000000953674,3.8199999332428,3.99000000953674,3.99000000953674,3300,0,0
2020-09-16 00:00:00-04:00,3.79999995231628,4,3.79999995231628,4,4,3300,0,0
2020-09-15 00:00:00-04:00,3.95000004768372,4,3.95000004768372,4,4,2400,0,0
2020-09-14 00:00:00-04:00,3.96000003814697,4,3.96000003814697,4,4,800,0,0
2020-09-11 00:00:00-04:00,3.95000004768372,3.97000002861023,3.72000002861023,3.97000002861023,3.97000002861023,5700,0,0
2020-09-10 00:00:00-04:00,4,4.09999990463257,4,4.09999990463257,4.09999990463257,7100,0,0
2020-09-09 00:00:00-04:00,3.5699999332428,4,3.5699999332428,4,4,18100,0,0
2020-09-08 00:00:00-04:00,3.40000009536743,3.59999990463257,3.40000009536743,3.59999990463257,3.59999990463257,19500,0,0
2020-09-04 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,400,0,0
2020-09-03 00:00:00-04:00,3.58999991416931,3.58999991416931,3.58999991416931,3.58999991416931,3.58999991416931,0,0,0
2020-09-02 00:00:00-04:00,3.5,3.58999991416931,3.5,3.58999991416931,3.58999991416931,2000,0,0
2020-09-01 00:00:00-04:00,3.5,3.59999990463257,3.5,3.59999990463257,3.59999990463257,1200,0,0
2020-08-31 00:00:00-04:00,3.15000009536743,3.70000004768372,3.15000009536743,3.70000004768372,3.70000004768372,26500,0,0
2020-08-28 00:00:00-04:00,3.76999998092651,3.76999998092651,3.70000004768372,3.70000004768372,3.70000004768372,1600,0,0
2020-08-27 00:00:00-04:00,3.65000009536743,3.65000009536743,3.65000009536743,3.65000009536743,3.65000009536743,0,0,0
2020-08-26 00:00:00-04:00,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,0,0,0.1
2020-08-25 00:00:00-04:00,3.40000009536743,3.70000004768372,3.40000009536743,3.70000004768372,3.70000004768372,2900,0,0
2020-08-24 00:00:00-04:00,3.29999995231628,3.5,3.29999995231628,3.5,3.5,10000,0,0
2020-08-21 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,150,0,0
2020-08-20 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-19 00:00:00-04:00,3.40000009536743,3.5,3.40000009536743,3.5,3.5,9050,0,0
2020-08-18 00:00:00-04:00,3.5,3.79999995231628,3.5,3.5,3.5,2250,0,0
2020-08-17 00:00:00-04:00,2.79999995231628,3.70000004768372,2.79999995231628,3.70000004768372,3.70000004768372,5050,0,0
2020-08-14 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-13 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-12 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-11 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-10 00:00:00-04:00,3.5,3.70000004768372,3.5,3.5,3.5,3300,0,0
2020-08-07 00:00:00-04:00,3.5,3.79999995231628,3.5,3.79999995231628,3.79999995231628,2500,0,0
2020-08-06 00:00:00-04:00,3.5,3.70000004768372,3.40000009536743,3.70000004768372,3.70000004768372,3000,0,0
2020-08-05 00:00:00-04:00,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,0,0,0
2020-08-04 00:00:00-04:00,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2020-09-30 00:00:00-04:00 4.40000009536743 4.44999980926514 4.01999998092651 4.44999980926514 4.44999980926514 22600 0 0
3 2020-09-29 00:00:00-04:00 4.3899998664856 4.40000009536743 4.13000011444092 4.30000019073486 4.30000019073486 10800 0 0
4 2020-09-28 00:00:00-04:00 4.09000015258789 4.25 4.09000015258789 4.25 4.25 8000 0 0
5 2020-09-25 00:00:00-04:00 3.95000004768372 4.09999990463257 3.95000004768372 4.05000019073486 4.05000019073486 13500 0 0
6 2020-09-24 00:00:00-04:00 3.84999990463257 4 3.84999990463257 4 4 8800 0 0
7 2020-09-23 00:00:00-04:00 3.99000000953674 4 3.99000000953674 4 4 5900 0 0
8 2020-09-22 00:00:00-04:00 3.90000009536743 4.09999990463257 3.84999990463257 4.09999990463257 4.09999990463257 3100 0 0
9 2020-09-21 00:00:00-04:00 4.09999990463257 4.09999990463257 4.09999990463257 4.09999990463257 4.09999990463257 1200 0 0
10 2020-09-18 00:00:00-04:00 3.92000007629395 4.09999990463257 3.92000007629395 4.09999990463257 4.09999990463257 27200 0 0
11 2020-09-17 00:00:00-04:00 3.90000009536743 3.99000000953674 3.8199999332428 3.99000000953674 3.99000000953674 3300 0 0
12 2020-09-16 00:00:00-04:00 3.79999995231628 4 3.79999995231628 4 4 3300 0 0
13 2020-09-15 00:00:00-04:00 3.95000004768372 4 3.95000004768372 4 4 2400 0 0
14 2020-09-14 00:00:00-04:00 3.96000003814697 4 3.96000003814697 4 4 800 0 0
15 2020-09-11 00:00:00-04:00 3.95000004768372 3.97000002861023 3.72000002861023 3.97000002861023 3.97000002861023 5700 0 0
16 2020-09-10 00:00:00-04:00 4 4.09999990463257 4 4.09999990463257 4.09999990463257 7100 0 0
17 2020-09-09 00:00:00-04:00 3.5699999332428 4 3.5699999332428 4 4 18100 0 0
18 2020-09-08 00:00:00-04:00 3.40000009536743 3.59999990463257 3.40000009536743 3.59999990463257 3.59999990463257 19500 0 0
19 2020-09-04 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 400 0 0
20 2020-09-03 00:00:00-04:00 3.58999991416931 3.58999991416931 3.58999991416931 3.58999991416931 3.58999991416931 0 0 0
21 2020-09-02 00:00:00-04:00 3.5 3.58999991416931 3.5 3.58999991416931 3.58999991416931 2000 0 0
22 2020-09-01 00:00:00-04:00 3.5 3.59999990463257 3.5 3.59999990463257 3.59999990463257 1200 0 0
23 2020-08-31 00:00:00-04:00 3.15000009536743 3.70000004768372 3.15000009536743 3.70000004768372 3.70000004768372 26500 0 0
24 2020-08-28 00:00:00-04:00 3.76999998092651 3.76999998092651 3.70000004768372 3.70000004768372 3.70000004768372 1600 0 0
25 2020-08-27 00:00:00-04:00 3.65000009536743 3.65000009536743 3.65000009536743 3.65000009536743 3.65000009536743 0 0 0
26 2020-08-26 00:00:00-04:00 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 0 0 0.1
27 2020-08-25 00:00:00-04:00 3.40000009536743 3.70000004768372 3.40000009536743 3.70000004768372 3.70000004768372 2900 0 0
28 2020-08-24 00:00:00-04:00 3.29999995231628 3.5 3.29999995231628 3.5 3.5 10000 0 0
29 2020-08-21 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 150 0 0
30 2020-08-20 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
31 2020-08-19 00:00:00-04:00 3.40000009536743 3.5 3.40000009536743 3.5 3.5 9050 0 0
32 2020-08-18 00:00:00-04:00 3.5 3.79999995231628 3.5 3.5 3.5 2250 0 0
33 2020-08-17 00:00:00-04:00 2.79999995231628 3.70000004768372 2.79999995231628 3.70000004768372 3.70000004768372 5050 0 0
34 2020-08-14 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
35 2020-08-13 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
36 2020-08-12 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
37 2020-08-11 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
38 2020-08-10 00:00:00-04:00 3.5 3.70000004768372 3.5 3.5 3.5 3300 0 0
39 2020-08-07 00:00:00-04:00 3.5 3.79999995231628 3.5 3.79999995231628 3.79999995231628 2500 0 0
40 2020-08-06 00:00:00-04:00 3.5 3.70000004768372 3.40000009536743 3.70000004768372 3.70000004768372 3000 0 0
41 2020-08-05 00:00:00-04:00 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 0 0 0
42 2020-08-04 00:00:00-04:00 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 0 0 0

View File

@@ -0,0 +1,42 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2020-09-30 00:00:00-04:00,4.40000009536743,4.44999980926514,4.01999998092651,4.44999980926514,4.44999980926514,22600,0,0
2020-09-29 00:00:00-04:00,4.3899998664856,4.40000009536743,4.13000011444092,4.30000019073486,4.30000019073486,10800,0,0
2020-09-28 00:00:00-04:00,4.09000015258789,4.25,4.09000015258789,4.25,4.25,8000,0,0
2020-09-25 00:00:00-04:00,3.95000004768372,4.09999990463257,3.95000004768372,4.05000019073486,4.05000019073486,13500,0,0
2020-09-24 00:00:00-04:00,3.84999990463257,4,3.84999990463257,4,4,8800,0,0
2020-09-23 00:00:00-04:00,3.99000000953674,4,3.99000000953674,4,4,5900,0,0
2020-09-22 00:00:00-04:00,3.90000009536743,4.09999990463257,3.84999990463257,4.09999990463257,4.09999990463257,3100,0,0
2020-09-21 00:00:00-04:00,4.09999990463257,4.09999990463257,4.09999990463257,4.09999990463257,4.09999990463257,1200,0,0
2020-09-18 00:00:00-04:00,3.92000007629395,4.09999990463257,3.92000007629395,4.09999990463257,4.09999990463257,27200,0,0
2020-09-17 00:00:00-04:00,3.90000009536743,3.99000000953674,3.8199999332428,3.99000000953674,3.99000000953674,3300,0,0
2020-09-16 00:00:00-04:00,3.79999995231628,4,3.79999995231628,4,4,3300,0,0
2020-09-15 00:00:00-04:00,3.95000004768372,4,3.95000004768372,4,4,2400,0,0
2020-09-14 00:00:00-04:00,3.96000003814697,4,3.96000003814697,4,4,800,0,0
2020-09-11 00:00:00-04:00,3.95000004768372,3.97000002861023,3.72000002861023,3.97000002861023,3.97000002861023,5700,0,0
2020-09-10 00:00:00-04:00,4,4.09999990463257,4,4.09999990463257,4.09999990463257,7100,0,0
2020-09-09 00:00:00-04:00,3.5699999332428,4,3.5699999332428,4,4,18100,0,0
2020-09-08 00:00:00-04:00,3.40000009536743,3.59999990463257,3.40000009536743,3.59999990463257,3.59999990463257,19500,0,0
2020-09-04 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,400,0,0
2020-09-03 00:00:00-04:00,3.58999991416931,3.58999991416931,3.58999991416931,3.58999991416931,3.58999991416931,0,0,0
2020-09-02 00:00:00-04:00,3.5,3.58999991416931,3.5,3.58999991416931,3.58999991416931,2000,0,0
2020-09-01 00:00:00-04:00,3.5,3.59999990463257,3.5,3.59999990463257,3.59999990463257,1200,0,0
2020-08-31 00:00:00-04:00,3.15000009536743,3.70000004768372,3.15000009536743,3.70000004768372,3.70000004768372,26500,0,0
2020-08-28 00:00:00-04:00,3.76999998092651,3.76999998092651,3.70000004768372,3.70000004768372,3.70000004768372,1600,0,0
2020-08-27 00:00:00-04:00,3.65000009536743,3.65000009536743,3.65000009536743,3.65000009536743,3.65000009536743,0,0,0
2020-08-26 00:00:00-04:00,0.370000004768372,0.370000004768372,0.370000004768372,0.370000004768372,0.370000004768372,0,0,0.1
2020-08-25 00:00:00-04:00,3.40000009536743,3.70000004768372,3.40000009536743,3.70000004768372,3.70000004768372,2900,0,0
2020-08-24 00:00:00-04:00,3.29999995231628,3.5,3.29999995231628,3.5,3.5,10000,0,0
2020-08-21 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,150,0,0
2020-08-20 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-19 00:00:00-04:00,3.40000009536743,3.5,3.40000009536743,3.5,3.5,9050,0,0
2020-08-18 00:00:00-04:00,3.5,3.79999995231628,3.5,3.5,3.5,2250,0,0
2020-08-17 00:00:00-04:00,2.79999995231628,3.70000004768372,2.79999995231628,3.70000004768372,3.70000004768372,5050,0,0
2020-08-14 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-13 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-12 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-11 00:00:00-04:00,3.5,3.5,3.5,3.5,3.5,0,0,0
2020-08-10 00:00:00-04:00,3.5,3.70000004768372,3.5,3.5,3.5,3300,0,0
2020-08-07 00:00:00-04:00,3.5,3.79999995231628,3.5,3.79999995231628,3.79999995231628,2500,0,0
2020-08-06 00:00:00-04:00,3.5,3.70000004768372,3.40000009536743,3.70000004768372,3.70000004768372,3000,0,0
2020-08-05 00:00:00-04:00,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,0,0,0
2020-08-04 00:00:00-04:00,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,3.70000004768372,0,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2020-09-30 00:00:00-04:00 4.40000009536743 4.44999980926514 4.01999998092651 4.44999980926514 4.44999980926514 22600 0 0
3 2020-09-29 00:00:00-04:00 4.3899998664856 4.40000009536743 4.13000011444092 4.30000019073486 4.30000019073486 10800 0 0
4 2020-09-28 00:00:00-04:00 4.09000015258789 4.25 4.09000015258789 4.25 4.25 8000 0 0
5 2020-09-25 00:00:00-04:00 3.95000004768372 4.09999990463257 3.95000004768372 4.05000019073486 4.05000019073486 13500 0 0
6 2020-09-24 00:00:00-04:00 3.84999990463257 4 3.84999990463257 4 4 8800 0 0
7 2020-09-23 00:00:00-04:00 3.99000000953674 4 3.99000000953674 4 4 5900 0 0
8 2020-09-22 00:00:00-04:00 3.90000009536743 4.09999990463257 3.84999990463257 4.09999990463257 4.09999990463257 3100 0 0
9 2020-09-21 00:00:00-04:00 4.09999990463257 4.09999990463257 4.09999990463257 4.09999990463257 4.09999990463257 1200 0 0
10 2020-09-18 00:00:00-04:00 3.92000007629395 4.09999990463257 3.92000007629395 4.09999990463257 4.09999990463257 27200 0 0
11 2020-09-17 00:00:00-04:00 3.90000009536743 3.99000000953674 3.8199999332428 3.99000000953674 3.99000000953674 3300 0 0
12 2020-09-16 00:00:00-04:00 3.79999995231628 4 3.79999995231628 4 4 3300 0 0
13 2020-09-15 00:00:00-04:00 3.95000004768372 4 3.95000004768372 4 4 2400 0 0
14 2020-09-14 00:00:00-04:00 3.96000003814697 4 3.96000003814697 4 4 800 0 0
15 2020-09-11 00:00:00-04:00 3.95000004768372 3.97000002861023 3.72000002861023 3.97000002861023 3.97000002861023 5700 0 0
16 2020-09-10 00:00:00-04:00 4 4.09999990463257 4 4.09999990463257 4.09999990463257 7100 0 0
17 2020-09-09 00:00:00-04:00 3.5699999332428 4 3.5699999332428 4 4 18100 0 0
18 2020-09-08 00:00:00-04:00 3.40000009536743 3.59999990463257 3.40000009536743 3.59999990463257 3.59999990463257 19500 0 0
19 2020-09-04 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 400 0 0
20 2020-09-03 00:00:00-04:00 3.58999991416931 3.58999991416931 3.58999991416931 3.58999991416931 3.58999991416931 0 0 0
21 2020-09-02 00:00:00-04:00 3.5 3.58999991416931 3.5 3.58999991416931 3.58999991416931 2000 0 0
22 2020-09-01 00:00:00-04:00 3.5 3.59999990463257 3.5 3.59999990463257 3.59999990463257 1200 0 0
23 2020-08-31 00:00:00-04:00 3.15000009536743 3.70000004768372 3.15000009536743 3.70000004768372 3.70000004768372 26500 0 0
24 2020-08-28 00:00:00-04:00 3.76999998092651 3.76999998092651 3.70000004768372 3.70000004768372 3.70000004768372 1600 0 0
25 2020-08-27 00:00:00-04:00 3.65000009536743 3.65000009536743 3.65000009536743 3.65000009536743 3.65000009536743 0 0 0
26 2020-08-26 00:00:00-04:00 0.370000004768372 0.370000004768372 0.370000004768372 0.370000004768372 0.370000004768372 0 0 0.1
27 2020-08-25 00:00:00-04:00 3.40000009536743 3.70000004768372 3.40000009536743 3.70000004768372 3.70000004768372 2900 0 0
28 2020-08-24 00:00:00-04:00 3.29999995231628 3.5 3.29999995231628 3.5 3.5 10000 0 0
29 2020-08-21 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 150 0 0
30 2020-08-20 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
31 2020-08-19 00:00:00-04:00 3.40000009536743 3.5 3.40000009536743 3.5 3.5 9050 0 0
32 2020-08-18 00:00:00-04:00 3.5 3.79999995231628 3.5 3.5 3.5 2250 0 0
33 2020-08-17 00:00:00-04:00 2.79999995231628 3.70000004768372 2.79999995231628 3.70000004768372 3.70000004768372 5050 0 0
34 2020-08-14 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
35 2020-08-13 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
36 2020-08-12 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
37 2020-08-11 00:00:00-04:00 3.5 3.5 3.5 3.5 3.5 0 0 0
38 2020-08-10 00:00:00-04:00 3.5 3.70000004768372 3.5 3.5 3.5 3300 0 0
39 2020-08-07 00:00:00-04:00 3.5 3.79999995231628 3.5 3.79999995231628 3.79999995231628 2500 0 0
40 2020-08-06 00:00:00-04:00 3.5 3.70000004768372 3.40000009536743 3.70000004768372 3.70000004768372 3000 0 0
41 2020-08-05 00:00:00-04:00 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 0 0 0
42 2020-08-04 00:00:00-04:00 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 3.70000004768372 0 0 0

View File

@@ -0,0 +1,17 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-08 00:00:00+02:00,24.8999996185303,24.9500007629395,24.1000003814697,24.75,24.75,7187,0,0
2023-05-09 00:00:00+02:00,25,25.5,23.1499996185303,24.1499996185303,24.1499996185303,22753,0,0
2023-05-10 00:00:00+02:00,24.1499996185303,24.1499996185303,22,22.9500007629395,22.9500007629395,62727,0,0
2023-05-11 00:00:00+02:00,22.9500007629395,25,22.9500007629395,23.3500003814697,23.3500003814697,19550,0,0
2023-05-12 00:00:00+02:00,23.3500003814697,24,22.1000003814697,23.8500003814697,23.8500003814697,17143,0,0
2023-05-15 00:00:00+02:00,23,25.7999992370605,22.5,23,23,43709,0,0
2023-05-16 00:00:00+02:00,22.75,24.0499992370605,22.5,22.75,22.75,16068,0,0
2023-05-17 00:00:00+02:00,23,23.8500003814697,22.1000003814697,23.6499996185303,23.6499996185303,19926,0,0
2023-05-19 00:00:00+02:00,23.6499996185303,23.8500003814697,22.1000003814697,22.2999992370605,22.2999992370605,41050,0,0
2023-05-22 00:00:00+02:00,22.0000004768372,24.1499996185303,21.5499997138977,22.7500009536743,22.7500009536743,34022,0,0
2023-05-23 00:00:00+02:00,22.75,22.8999996185303,21.75,22.5,22.5,13992,0,0
2023-05-24 00:00:00+02:00,21,24,21,22.0100002288818,22.0100002288818,18306,0,0.1
2023-05-25 00:00:00+02:00,21.5699996948242,22.8899993896484,20,21.1599998474121,21.1599998474121,35398,0,0
2023-05-26 00:00:00+02:00,21.1599998474121,22.4950008392334,20.5,21.0949993133545,21.0949993133545,8039,0,0
2023-05-29 00:00:00+02:00,22.1000003814697,22.1000003814697,20.25,20.75,20.75,17786,0,0
2023-05-30 00:00:00+02:00,20.75,21.6499996185303,20.1499996185303,20.4500007629395,20.4500007629395,10709,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-08 00:00:00+02:00 24.8999996185303 24.9500007629395 24.1000003814697 24.75 24.75 7187 0 0
3 2023-05-09 00:00:00+02:00 25 25.5 23.1499996185303 24.1499996185303 24.1499996185303 22753 0 0
4 2023-05-10 00:00:00+02:00 24.1499996185303 24.1499996185303 22 22.9500007629395 22.9500007629395 62727 0 0
5 2023-05-11 00:00:00+02:00 22.9500007629395 25 22.9500007629395 23.3500003814697 23.3500003814697 19550 0 0
6 2023-05-12 00:00:00+02:00 23.3500003814697 24 22.1000003814697 23.8500003814697 23.8500003814697 17143 0 0
7 2023-05-15 00:00:00+02:00 23 25.7999992370605 22.5 23 23 43709 0 0
8 2023-05-16 00:00:00+02:00 22.75 24.0499992370605 22.5 22.75 22.75 16068 0 0
9 2023-05-17 00:00:00+02:00 23 23.8500003814697 22.1000003814697 23.6499996185303 23.6499996185303 19926 0 0
10 2023-05-19 00:00:00+02:00 23.6499996185303 23.8500003814697 22.1000003814697 22.2999992370605 22.2999992370605 41050 0 0
11 2023-05-22 00:00:00+02:00 22.0000004768372 24.1499996185303 21.5499997138977 22.7500009536743 22.7500009536743 34022 0 0
12 2023-05-23 00:00:00+02:00 22.75 22.8999996185303 21.75 22.5 22.5 13992 0 0
13 2023-05-24 00:00:00+02:00 21 24 21 22.0100002288818 22.0100002288818 18306 0 0.1
14 2023-05-25 00:00:00+02:00 21.5699996948242 22.8899993896484 20 21.1599998474121 21.1599998474121 35398 0 0
15 2023-05-26 00:00:00+02:00 21.1599998474121 22.4950008392334 20.5 21.0949993133545 21.0949993133545 8039 0 0
16 2023-05-29 00:00:00+02:00 22.1000003814697 22.1000003814697 20.25 20.75 20.75 17786 0 0
17 2023-05-30 00:00:00+02:00 20.75 21.6499996185303 20.1499996185303 20.4500007629395 20.4500007629395 10709 0 0

View File

@@ -0,0 +1,17 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-05-08 00:00:00+02:00,24.899999618530273,24.950000762939453,24.100000381469727,24.75,24.75,7187,0.0,0.0
2023-05-09 00:00:00+02:00,25.0,25.5,23.149999618530273,24.149999618530273,24.149999618530273,22753,0.0,0.0
2023-05-10 00:00:00+02:00,24.149999618530273,24.149999618530273,22.0,22.950000762939453,22.950000762939453,62727,0.0,0.0
2023-05-11 00:00:00+02:00,22.950000762939453,25.0,22.950000762939453,23.350000381469727,23.350000381469727,19550,0.0,0.0
2023-05-12 00:00:00+02:00,23.350000381469727,24.0,22.100000381469727,23.850000381469727,23.850000381469727,17143,0.0,0.0
2023-05-15 00:00:00+02:00,23.0,25.799999237060547,22.5,23.0,23.0,43709,0.0,0.0
2023-05-16 00:00:00+02:00,22.75,24.049999237060547,22.5,22.75,22.75,16068,0.0,0.0
2023-05-17 00:00:00+02:00,23.0,23.850000381469727,22.100000381469727,23.649999618530273,23.649999618530273,19926,0.0,0.0
2023-05-19 00:00:00+02:00,23.649999618530273,23.850000381469727,22.100000381469727,22.299999237060547,22.299999237060547,41050,0.0,0.0
2023-05-22 00:00:00+02:00,2.200000047683716,2.4149999618530273,2.1549999713897705,2.2750000953674316,2.2750000953674316,340215,0.0,0.0
2023-05-23 00:00:00+02:00,22.75,22.899999618530273,21.75,22.5,22.5,13992,0.0,0.0
2023-05-24 00:00:00+02:00,21.0,24.0,21.0,22.010000228881836,22.010000228881836,18306,0.0,0.1
2023-05-25 00:00:00+02:00,21.56999969482422,22.889999389648438,20.0,21.15999984741211,21.15999984741211,35398,0.0,0.0
2023-05-26 00:00:00+02:00,21.15999984741211,22.4950008392334,20.5,21.094999313354492,21.094999313354492,8039,0.0,0.0
2023-05-29 00:00:00+02:00,22.100000381469727,22.100000381469727,20.25,20.75,20.75,17786,0.0,0.0
2023-05-30 00:00:00+02:00,20.75,21.649999618530273,20.149999618530273,20.450000762939453,20.450000762939453,10709,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-05-08 00:00:00+02:00 24.899999618530273 24.950000762939453 24.100000381469727 24.75 24.75 7187 0.0 0.0
3 2023-05-09 00:00:00+02:00 25.0 25.5 23.149999618530273 24.149999618530273 24.149999618530273 22753 0.0 0.0
4 2023-05-10 00:00:00+02:00 24.149999618530273 24.149999618530273 22.0 22.950000762939453 22.950000762939453 62727 0.0 0.0
5 2023-05-11 00:00:00+02:00 22.950000762939453 25.0 22.950000762939453 23.350000381469727 23.350000381469727 19550 0.0 0.0
6 2023-05-12 00:00:00+02:00 23.350000381469727 24.0 22.100000381469727 23.850000381469727 23.850000381469727 17143 0.0 0.0
7 2023-05-15 00:00:00+02:00 23.0 25.799999237060547 22.5 23.0 23.0 43709 0.0 0.0
8 2023-05-16 00:00:00+02:00 22.75 24.049999237060547 22.5 22.75 22.75 16068 0.0 0.0
9 2023-05-17 00:00:00+02:00 23.0 23.850000381469727 22.100000381469727 23.649999618530273 23.649999618530273 19926 0.0 0.0
10 2023-05-19 00:00:00+02:00 23.649999618530273 23.850000381469727 22.100000381469727 22.299999237060547 22.299999237060547 41050 0.0 0.0
11 2023-05-22 00:00:00+02:00 2.200000047683716 2.4149999618530273 2.1549999713897705 2.2750000953674316 2.2750000953674316 340215 0.0 0.0
12 2023-05-23 00:00:00+02:00 22.75 22.899999618530273 21.75 22.5 22.5 13992 0.0 0.0
13 2023-05-24 00:00:00+02:00 21.0 24.0 21.0 22.010000228881836 22.010000228881836 18306 0.0 0.1
14 2023-05-25 00:00:00+02:00 21.56999969482422 22.889999389648438 20.0 21.15999984741211 21.15999984741211 35398 0.0 0.0
15 2023-05-26 00:00:00+02:00 21.15999984741211 22.4950008392334 20.5 21.094999313354492 21.094999313354492 8039 0.0 0.0
16 2023-05-29 00:00:00+02:00 22.100000381469727 22.100000381469727 20.25 20.75 20.75 17786 0.0 0.0
17 2023-05-30 00:00:00+02:00 20.75 21.649999618530273 20.149999618530273 20.450000762939453 20.450000762939453 10709 0.0 0.0

View File

@@ -0,0 +1,23 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-06-01 00:00:00+02:00,5.72999992370606,5.78199996948242,5.3939998626709,5.3939998626709,5.3939998626709,3095860,0,0
2022-06-02 00:00:00+02:00,5.38600006103516,5.38600006103516,5.26800003051758,5.2939998626709,5.2939998626709,1662880,0,0
2022-06-03 00:00:00+02:00,5.34599990844727,5.34599990844727,5.15800018310547,5.16800003051758,5.16800003051758,1698900,0,0
2022-06-06 00:00:00+02:00,5.16800003051758,5.25200004577637,5.13800010681152,5.18800010681152,5.18800010681152,1074910,0,0
2022-06-07 00:00:00+02:00,5.21800003051758,5.22200012207031,5.07400016784668,5.1560001373291,5.1560001373291,1850680,0,0
2022-06-08 00:00:00+02:00,5.1560001373291,5.17599983215332,5.07200012207031,5.10200004577637,5.10200004577637,1140360,0,0
2022-06-09 00:00:00+02:00,5.09799995422363,5.09799995422363,4.87599983215332,4.8939998626709,4.8939998626709,2025480,0,0
2022-06-10 00:00:00+02:00,4.87999992370606,4.87999992370606,4.50400009155274,4.50400009155274,4.50400009155274,2982730,0,0
2022-06-13 00:00:00+02:00,4.3,4.37599983215332,3.83600006103516,3.83600006103516,3.83600006103516,4568210,0,0.1
2022-06-14 00:00:00+02:00,3.87750015258789,4.15999984741211,3.85200004577637,3.9439998626709,3.9439998626709,5354500,0,0
2022-06-15 00:00:00+02:00,4.03400001525879,4.16450004577637,3.73050003051758,3.73050003051758,3.73050003051758,6662610,0,0
2022-06-16 00:00:00+02:00,3.73050003051758,3.98499984741211,3.72400016784668,3.82550010681152,3.82550010681152,13379960,0,0
2022-06-17 00:00:00+02:00,3.8,4.29949989318848,3.75,4.29949989318848,4.29949989318848,12844160,0,0
2022-06-20 00:00:00+02:00,2.19422197341919,2.2295401096344,2.13992595672607,2.2295401096344,2.2295401096344,12364104,0,0
2022-06-21 00:00:00+02:00,2.24719905853272,2.28515291213989,2.19712090492249,2.21557092666626,2.21557092666626,8434013,0,0
2022-06-22 00:00:00+02:00,1.98679196834564,2.00365996360779,1.73798203468323,1.73798203468323,1.73798203468323,26496542,0,0
2022-06-23 00:00:00+02:00,1.62411904335022,1.68526804447174,1.37320005893707,1.59776198863983,1.59776198863983,48720201,0,0
2022-06-24 00:00:00+02:00,1.47599303722382,1.54610300064087,1.1739410161972,1.24932205677032,1.24932205677032,56877192,0,0
2022-06-27 00:00:00+02:00,1.49899995326996,1.79849994182587,1.49899995326996,1.79849994182587,1.79849994182587,460673,0,0
2022-06-28 00:00:00+02:00,2.15799999237061,3.05100011825562,2.12599992752075,3.05100011825562,3.05100011825562,3058635,0,0
2022-06-29 00:00:00+02:00,2.90000009536743,3.73799991607666,2.85899996757507,3.26399993896484,3.26399993896484,6516761,0,0
2022-06-30 00:00:00+02:00,3.24900007247925,3.28099989891052,2.5,2.5550000667572,2.5550000667572,4805984,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-06-01 00:00:00+02:00 5.72999992370606 5.78199996948242 5.3939998626709 5.3939998626709 5.3939998626709 3095860 0 0
3 2022-06-02 00:00:00+02:00 5.38600006103516 5.38600006103516 5.26800003051758 5.2939998626709 5.2939998626709 1662880 0 0
4 2022-06-03 00:00:00+02:00 5.34599990844727 5.34599990844727 5.15800018310547 5.16800003051758 5.16800003051758 1698900 0 0
5 2022-06-06 00:00:00+02:00 5.16800003051758 5.25200004577637 5.13800010681152 5.18800010681152 5.18800010681152 1074910 0 0
6 2022-06-07 00:00:00+02:00 5.21800003051758 5.22200012207031 5.07400016784668 5.1560001373291 5.1560001373291 1850680 0 0
7 2022-06-08 00:00:00+02:00 5.1560001373291 5.17599983215332 5.07200012207031 5.10200004577637 5.10200004577637 1140360 0 0
8 2022-06-09 00:00:00+02:00 5.09799995422363 5.09799995422363 4.87599983215332 4.8939998626709 4.8939998626709 2025480 0 0
9 2022-06-10 00:00:00+02:00 4.87999992370606 4.87999992370606 4.50400009155274 4.50400009155274 4.50400009155274 2982730 0 0
10 2022-06-13 00:00:00+02:00 4.3 4.37599983215332 3.83600006103516 3.83600006103516 3.83600006103516 4568210 0 0.1
11 2022-06-14 00:00:00+02:00 3.87750015258789 4.15999984741211 3.85200004577637 3.9439998626709 3.9439998626709 5354500 0 0
12 2022-06-15 00:00:00+02:00 4.03400001525879 4.16450004577637 3.73050003051758 3.73050003051758 3.73050003051758 6662610 0 0
13 2022-06-16 00:00:00+02:00 3.73050003051758 3.98499984741211 3.72400016784668 3.82550010681152 3.82550010681152 13379960 0 0
14 2022-06-17 00:00:00+02:00 3.8 4.29949989318848 3.75 4.29949989318848 4.29949989318848 12844160 0 0
15 2022-06-20 00:00:00+02:00 2.19422197341919 2.2295401096344 2.13992595672607 2.2295401096344 2.2295401096344 12364104 0 0
16 2022-06-21 00:00:00+02:00 2.24719905853272 2.28515291213989 2.19712090492249 2.21557092666626 2.21557092666626 8434013 0 0
17 2022-06-22 00:00:00+02:00 1.98679196834564 2.00365996360779 1.73798203468323 1.73798203468323 1.73798203468323 26496542 0 0
18 2022-06-23 00:00:00+02:00 1.62411904335022 1.68526804447174 1.37320005893707 1.59776198863983 1.59776198863983 48720201 0 0
19 2022-06-24 00:00:00+02:00 1.47599303722382 1.54610300064087 1.1739410161972 1.24932205677032 1.24932205677032 56877192 0 0
20 2022-06-27 00:00:00+02:00 1.49899995326996 1.79849994182587 1.49899995326996 1.79849994182587 1.79849994182587 460673 0 0
21 2022-06-28 00:00:00+02:00 2.15799999237061 3.05100011825562 2.12599992752075 3.05100011825562 3.05100011825562 3058635 0 0
22 2022-06-29 00:00:00+02:00 2.90000009536743 3.73799991607666 2.85899996757507 3.26399993896484 3.26399993896484 6516761 0 0
23 2022-06-30 00:00:00+02:00 3.24900007247925 3.28099989891052 2.5 2.5550000667572 2.5550000667572 4805984 0 0

View File

@@ -0,0 +1,23 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2022-06-01 00:00:00+02:00,57.29999923706055,57.81999969482422,53.939998626708984,53.939998626708984,53.939998626708984,309586,0.0,0.0
2022-06-02 00:00:00+02:00,53.86000061035156,53.86000061035156,52.68000030517578,52.939998626708984,52.939998626708984,166288,0.0,0.0
2022-06-03 00:00:00+02:00,53.459999084472656,53.459999084472656,51.58000183105469,51.68000030517578,51.68000030517578,169890,0.0,0.0
2022-06-06 00:00:00+02:00,51.68000030517578,52.52000045776367,51.380001068115234,51.880001068115234,51.880001068115234,107491,0.0,0.0
2022-06-07 00:00:00+02:00,52.18000030517578,52.220001220703125,50.7400016784668,51.560001373291016,51.560001373291016,185068,0.0,0.0
2022-06-08 00:00:00+02:00,51.560001373291016,51.7599983215332,50.720001220703125,51.02000045776367,51.02000045776367,114036,0.0,0.0
2022-06-09 00:00:00+02:00,50.97999954223633,50.97999954223633,48.7599983215332,48.939998626708984,48.939998626708984,202548,0.0,0.0
2022-06-10 00:00:00+02:00,48.79999923706055,48.79999923706055,45.040000915527344,45.040000915527344,45.040000915527344,298273,0.0,0.0
2022-06-13 00:00:00+02:00,43.0,43.7599983215332,38.36000061035156,38.36000061035156,38.36000061035156,456821,0.0,0.1
2022-06-14 00:00:00+02:00,38.775001525878906,41.599998474121094,38.52000045776367,39.439998626708984,39.439998626708984,535450,0.0,0.0
2022-06-15 00:00:00+02:00,40.34000015258789,41.64500045776367,37.30500030517578,37.30500030517578,37.30500030517578,666261,0.0,0.0
2022-06-16 00:00:00+02:00,37.30500030517578,39.849998474121094,37.2400016784668,38.255001068115234,38.255001068115234,1337996,0.0,0.0
2022-06-17 00:00:00+02:00,38.0,42.994998931884766,37.5,42.994998931884766,42.994998931884766,1284416,0.0,0.0
2022-06-20 00:00:00+02:00,2.1942219734191895,2.2295401096343994,2.139925956726074,2.2295401096343994,2.2295401096343994,12364104,0.0,0.0
2022-06-21 00:00:00+02:00,2.247199058532715,2.2851529121398926,2.1971209049224854,2.2155709266662598,2.2155709266662598,8434013,0.0,0.0
2022-06-22 00:00:00+02:00,1.986791968345642,2.003659963607788,1.7379820346832275,1.7379820346832275,1.7379820346832275,26496542,0.0,0.0
2022-06-23 00:00:00+02:00,1.6241190433502197,1.6852680444717407,1.3732000589370728,1.5977619886398315,1.5977619886398315,48720201,0.0,0.0
2022-06-24 00:00:00+02:00,1.475993037223816,1.5461030006408691,1.1739410161972046,1.2493220567703247,1.2493220567703247,56877192,0.0,0.0
2022-06-27 00:00:00+02:00,1.4989999532699585,1.7984999418258667,1.4989999532699585,1.7984999418258667,1.7984999418258667,460673,0.0,0.0
2022-06-28 00:00:00+02:00,2.1579999923706055,3.0510001182556152,2.125999927520752,3.0510001182556152,3.0510001182556152,3058635,0.0,0.0
2022-06-29 00:00:00+02:00,2.9000000953674316,3.73799991607666,2.8589999675750732,3.2639999389648438,3.2639999389648438,6516761,0.0,0.0
2022-06-30 00:00:00+02:00,3.249000072479248,3.2809998989105225,2.5,2.555000066757202,2.555000066757202,4805984,0.0,0.0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2022-06-01 00:00:00+02:00 57.29999923706055 57.81999969482422 53.939998626708984 53.939998626708984 53.939998626708984 309586 0.0 0.0
3 2022-06-02 00:00:00+02:00 53.86000061035156 53.86000061035156 52.68000030517578 52.939998626708984 52.939998626708984 166288 0.0 0.0
4 2022-06-03 00:00:00+02:00 53.459999084472656 53.459999084472656 51.58000183105469 51.68000030517578 51.68000030517578 169890 0.0 0.0
5 2022-06-06 00:00:00+02:00 51.68000030517578 52.52000045776367 51.380001068115234 51.880001068115234 51.880001068115234 107491 0.0 0.0
6 2022-06-07 00:00:00+02:00 52.18000030517578 52.220001220703125 50.7400016784668 51.560001373291016 51.560001373291016 185068 0.0 0.0
7 2022-06-08 00:00:00+02:00 51.560001373291016 51.7599983215332 50.720001220703125 51.02000045776367 51.02000045776367 114036 0.0 0.0
8 2022-06-09 00:00:00+02:00 50.97999954223633 50.97999954223633 48.7599983215332 48.939998626708984 48.939998626708984 202548 0.0 0.0
9 2022-06-10 00:00:00+02:00 48.79999923706055 48.79999923706055 45.040000915527344 45.040000915527344 45.040000915527344 298273 0.0 0.0
10 2022-06-13 00:00:00+02:00 43.0 43.7599983215332 38.36000061035156 38.36000061035156 38.36000061035156 456821 0.0 0.1
11 2022-06-14 00:00:00+02:00 38.775001525878906 41.599998474121094 38.52000045776367 39.439998626708984 39.439998626708984 535450 0.0 0.0
12 2022-06-15 00:00:00+02:00 40.34000015258789 41.64500045776367 37.30500030517578 37.30500030517578 37.30500030517578 666261 0.0 0.0
13 2022-06-16 00:00:00+02:00 37.30500030517578 39.849998474121094 37.2400016784668 38.255001068115234 38.255001068115234 1337996 0.0 0.0
14 2022-06-17 00:00:00+02:00 38.0 42.994998931884766 37.5 42.994998931884766 42.994998931884766 1284416 0.0 0.0
15 2022-06-20 00:00:00+02:00 2.1942219734191895 2.2295401096343994 2.139925956726074 2.2295401096343994 2.2295401096343994 12364104 0.0 0.0
16 2022-06-21 00:00:00+02:00 2.247199058532715 2.2851529121398926 2.1971209049224854 2.2155709266662598 2.2155709266662598 8434013 0.0 0.0
17 2022-06-22 00:00:00+02:00 1.986791968345642 2.003659963607788 1.7379820346832275 1.7379820346832275 1.7379820346832275 26496542 0.0 0.0
18 2022-06-23 00:00:00+02:00 1.6241190433502197 1.6852680444717407 1.3732000589370728 1.5977619886398315 1.5977619886398315 48720201 0.0 0.0
19 2022-06-24 00:00:00+02:00 1.475993037223816 1.5461030006408691 1.1739410161972046 1.2493220567703247 1.2493220567703247 56877192 0.0 0.0
20 2022-06-27 00:00:00+02:00 1.4989999532699585 1.7984999418258667 1.4989999532699585 1.7984999418258667 1.7984999418258667 460673 0.0 0.0
21 2022-06-28 00:00:00+02:00 2.1579999923706055 3.0510001182556152 2.125999927520752 3.0510001182556152 3.0510001182556152 3058635 0.0 0.0
22 2022-06-29 00:00:00+02:00 2.9000000953674316 3.73799991607666 2.8589999675750732 3.2639999389648438 3.2639999389648438 6516761 0.0 0.0
23 2022-06-30 00:00:00+02:00 3.249000072479248 3.2809998989105225 2.5 2.555000066757202 2.555000066757202 4805984 0.0 0.0

View File

@@ -0,0 +1,30 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-06-09 00:00:00+02:00,34.7000,34.7100,33.2400,33.6200,33.6200,7148409,0,0
2023-06-08 00:00:00+02:00,34.9000,34.9900,34.0400,34.3600,34.3600,10406999,0,0
2023-06-07 00:00:00+02:00,34.5500,35.6400,34.3200,35.0900,35.0900,10118918,0,0
2023-06-06 00:00:00+02:00,34.5000,34.8200,34.0500,34.4600,34.4600,9109709,0,0
2023-06-05 00:00:00+02:00,35.0000,35.3000,34.2000,34.7000,34.7000,8791993,0,0
2023-06-02 00:00:00+02:00,35.6900,36.1800,34.6000,34.9700,34.9700,8844549,0,0
2023-06-01 00:00:00+02:00,35.2300,35.3800,34.2400,35.3500,35.3500,6721030,0,0
2023-05-31 00:00:00+02:00,34.8,35.48,34.26,35.01,35.01,32605833,0,0
2023-05-30 00:00:00+02:00,34.39,35.37,33.85,34.23,34.23,8970804,0,0
2023-05-29 00:00:00+02:00,34.66,35.06,34.02,34.32,34.32,3912803,0,0
2023-05-26 00:00:00+02:00,34.75,35.99,34.33,34.53,34.53,6744718,0,0
2023-05-25 00:00:00+02:00,35.4,36.09,34.63,35.07,35.07,16900221,0,0
2023-05-24 00:00:00+02:00,36.2,36.5,35.26,35.4,35.4,9049505,0,0
2023-05-23 00:00:00+02:00,36.9,36.67,35.56,36.1,36.1,10797373,0,0
2023-05-22 00:00:00+02:00,37.05,37.36,36.09,36.61,36.61,7132641,0,0
2023-05-19 00:00:00+02:00,36.2,37.15,36.25,36.9,36.9,12648518,0,0
2023-05-18 00:00:00+02:00,36.57,36.99,35.84,36.46,36.46,10674542,0,0
2023-05-17 00:00:00+02:00,36.87,37.31,36.56,36.71,36.71,9892791,0,0
2023-05-16 00:00:00+02:00,37.15,37.73,36.96,37.03,37.03,4706789,0,0
2023-05-15 00:00:00+02:00,37.74,38.05,36.96,37.27,37.27,7890969,0,0
2023-05-12 00:00:00+02:00,37.5,38.44,36.71,37.74,37.74,8724303,0,0
2023-05-11 00:00:00+02:00,38.8,38.88,37.01,37.32,37.32,14371855,0,0
2023-05-10 00:00:00+02:00,38.93,38.8,36.42,38.1,38.1,30393389,0,0
2023-05-09 00:00:00+02:00,44.41,44.41,39.39,39.66,39.66,19833428,0,0
2023-05-08 00:00:00+02:00,44.63,45.78,44.56,44.71,44.71,11092519,0,0
2023-05-05 00:00:00+02:00,42.99,44.9,42.87,44.58,44.58,28539048,0,0
2023-05-04 00:00:00+02:00,41.49,43.3,41.23,42.83,42.83,15506868,0,0
2023-05-03 00:00:00+02:00,39.75,40.98,39.68,40.95,40.95,14657028,0,0
2023-05-02 00:00:00+02:00,40.37,40.32,39.17,39.65,39.65,11818133,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-06-09 00:00:00+02:00 34.7000 34.7100 33.2400 33.6200 33.6200 7148409 0 0
3 2023-06-08 00:00:00+02:00 34.9000 34.9900 34.0400 34.3600 34.3600 10406999 0 0
4 2023-06-07 00:00:00+02:00 34.5500 35.6400 34.3200 35.0900 35.0900 10118918 0 0
5 2023-06-06 00:00:00+02:00 34.5000 34.8200 34.0500 34.4600 34.4600 9109709 0 0
6 2023-06-05 00:00:00+02:00 35.0000 35.3000 34.2000 34.7000 34.7000 8791993 0 0
7 2023-06-02 00:00:00+02:00 35.6900 36.1800 34.6000 34.9700 34.9700 8844549 0 0
8 2023-06-01 00:00:00+02:00 35.2300 35.3800 34.2400 35.3500 35.3500 6721030 0 0
9 2023-05-31 00:00:00+02:00 34.8 35.48 34.26 35.01 35.01 32605833 0 0
10 2023-05-30 00:00:00+02:00 34.39 35.37 33.85 34.23 34.23 8970804 0 0
11 2023-05-29 00:00:00+02:00 34.66 35.06 34.02 34.32 34.32 3912803 0 0
12 2023-05-26 00:00:00+02:00 34.75 35.99 34.33 34.53 34.53 6744718 0 0
13 2023-05-25 00:00:00+02:00 35.4 36.09 34.63 35.07 35.07 16900221 0 0
14 2023-05-24 00:00:00+02:00 36.2 36.5 35.26 35.4 35.4 9049505 0 0
15 2023-05-23 00:00:00+02:00 36.9 36.67 35.56 36.1 36.1 10797373 0 0
16 2023-05-22 00:00:00+02:00 37.05 37.36 36.09 36.61 36.61 7132641 0 0
17 2023-05-19 00:00:00+02:00 36.2 37.15 36.25 36.9 36.9 12648518 0 0
18 2023-05-18 00:00:00+02:00 36.57 36.99 35.84 36.46 36.46 10674542 0 0
19 2023-05-17 00:00:00+02:00 36.87 37.31 36.56 36.71 36.71 9892791 0 0
20 2023-05-16 00:00:00+02:00 37.15 37.73 36.96 37.03 37.03 4706789 0 0
21 2023-05-15 00:00:00+02:00 37.74 38.05 36.96 37.27 37.27 7890969 0 0
22 2023-05-12 00:00:00+02:00 37.5 38.44 36.71 37.74 37.74 8724303 0 0
23 2023-05-11 00:00:00+02:00 38.8 38.88 37.01 37.32 37.32 14371855 0 0
24 2023-05-10 00:00:00+02:00 38.93 38.8 36.42 38.1 38.1 30393389 0 0
25 2023-05-09 00:00:00+02:00 44.41 44.41 39.39 39.66 39.66 19833428 0 0
26 2023-05-08 00:00:00+02:00 44.63 45.78 44.56 44.71 44.71 11092519 0 0
27 2023-05-05 00:00:00+02:00 42.99 44.9 42.87 44.58 44.58 28539048 0 0
28 2023-05-04 00:00:00+02:00 41.49 43.3 41.23 42.83 42.83 15506868 0 0
29 2023-05-03 00:00:00+02:00 39.75 40.98 39.68 40.95 40.95 14657028 0 0
30 2023-05-02 00:00:00+02:00 40.37 40.32 39.17 39.65 39.65 11818133 0 0

View File

@@ -0,0 +1,30 @@
Date,Open,High,Low,Close,Adj Close,Volume,Dividends,Stock Splits
2023-06-09 00:00:00+02:00,34.700001,34.709999,33.240002,33.619999,33.619999,7148409,0,0
2023-06-08 00:00:00+02:00,34.900002,34.990002,34.040001,34.360001,34.360001,10406999,0,0
2023-06-07 00:00:00+02:00,34.549999,35.639999,34.320000,35.090000,35.090000,10118918,0,0
2023-06-06 00:00:00+02:00,34.500000,34.820000,34.049999,34.459999,34.459999,9109709,0,0
2023-06-05 00:00:00+02:00,35.000000,35.299999,34.200001,34.700001,34.700001,8791993,0,0
2023-06-02 00:00:00+02:00,35.689999,36.180000,34.599998,34.970001,34.970001,8844549,0,0
2023-06-01 00:00:00+02:00,35.230000,35.380001,34.240002,35.349998,35.349998,6721030,0,0
2023-05-31 00:00:00+02:00,3480,3548,3426,3501,3501,32605833,0,0
2023-05-30 00:00:00+02:00,3439,3537,3385,3423,3423,8970804,0,0
2023-05-29 00:00:00+02:00,3466,3506,3402,3432,3432,3912803,0,0
2023-05-26 00:00:00+02:00,3475,3599,3433,3453,3453,6744718,0,0
2023-05-25 00:00:00+02:00,3540,3609,3463,3507,3507,16900221,0,0
2023-05-24 00:00:00+02:00,3620,3650,3526,3540,3540,9049505,0,0
2023-05-23 00:00:00+02:00,3690,3667,3556,3610,3610,10797373,0,0
2023-05-22 00:00:00+02:00,3705,3736,3609,3661,3661,7132641,0,0
2023-05-19 00:00:00+02:00,3620,3715,3625,3690,3690,12648518,0,0
2023-05-18 00:00:00+02:00,3657,3699,3584,3646,3646,10674542,0,0
2023-05-17 00:00:00+02:00,3687,3731,3656,3671,3671,9892791,0,0
2023-05-16 00:00:00+02:00,3715,3773,3696,3703,3703,4706789,0,0
2023-05-15 00:00:00+02:00,3774,3805,3696,3727,3727,7890969,0,0
2023-05-12 00:00:00+02:00,3750,3844,3671,3774,3774,8724303,0,0
2023-05-11 00:00:00+02:00,3880,3888,3701,3732,3732,14371855,0,0
2023-05-10 00:00:00+02:00,3893,3880,3642,3810,3810,30393389,0,0
2023-05-09 00:00:00+02:00,4441,4441,3939,3966,3966,19833428,0,0
2023-05-08 00:00:00+02:00,4463,4578,4456,4471,4471,11092519,0,0
2023-05-05 00:00:00+02:00,4299,4490,4287,4458,4458,28539048,0,0
2023-05-04 00:00:00+02:00,4149,4330,4123,4283,4283,15506868,0,0
2023-05-03 00:00:00+02:00,3975,4098,3968,4095,4095,14657028,0,0
2023-05-02 00:00:00+02:00,4037,4032,3917,3965,3965,11818133,0,0
1 Date Open High Low Close Adj Close Volume Dividends Stock Splits
2 2023-06-09 00:00:00+02:00 34.700001 34.709999 33.240002 33.619999 33.619999 7148409 0 0
3 2023-06-08 00:00:00+02:00 34.900002 34.990002 34.040001 34.360001 34.360001 10406999 0 0
4 2023-06-07 00:00:00+02:00 34.549999 35.639999 34.320000 35.090000 35.090000 10118918 0 0
5 2023-06-06 00:00:00+02:00 34.500000 34.820000 34.049999 34.459999 34.459999 9109709 0 0
6 2023-06-05 00:00:00+02:00 35.000000 35.299999 34.200001 34.700001 34.700001 8791993 0 0
7 2023-06-02 00:00:00+02:00 35.689999 36.180000 34.599998 34.970001 34.970001 8844549 0 0
8 2023-06-01 00:00:00+02:00 35.230000 35.380001 34.240002 35.349998 35.349998 6721030 0 0
9 2023-05-31 00:00:00+02:00 3480 3548 3426 3501 3501 32605833 0 0
10 2023-05-30 00:00:00+02:00 3439 3537 3385 3423 3423 8970804 0 0
11 2023-05-29 00:00:00+02:00 3466 3506 3402 3432 3432 3912803 0 0
12 2023-05-26 00:00:00+02:00 3475 3599 3433 3453 3453 6744718 0 0
13 2023-05-25 00:00:00+02:00 3540 3609 3463 3507 3507 16900221 0 0
14 2023-05-24 00:00:00+02:00 3620 3650 3526 3540 3540 9049505 0 0
15 2023-05-23 00:00:00+02:00 3690 3667 3556 3610 3610 10797373 0 0
16 2023-05-22 00:00:00+02:00 3705 3736 3609 3661 3661 7132641 0 0
17 2023-05-19 00:00:00+02:00 3620 3715 3625 3690 3690 12648518 0 0
18 2023-05-18 00:00:00+02:00 3657 3699 3584 3646 3646 10674542 0 0
19 2023-05-17 00:00:00+02:00 3687 3731 3656 3671 3671 9892791 0 0
20 2023-05-16 00:00:00+02:00 3715 3773 3696 3703 3703 4706789 0 0
21 2023-05-15 00:00:00+02:00 3774 3805 3696 3727 3727 7890969 0 0
22 2023-05-12 00:00:00+02:00 3750 3844 3671 3774 3774 8724303 0 0
23 2023-05-11 00:00:00+02:00 3880 3888 3701 3732 3732 14371855 0 0
24 2023-05-10 00:00:00+02:00 3893 3880 3642 3810 3810 30393389 0 0
25 2023-05-09 00:00:00+02:00 4441 4441 3939 3966 3966 19833428 0 0
26 2023-05-08 00:00:00+02:00 4463 4578 4456 4471 4471 11092519 0 0
27 2023-05-05 00:00:00+02:00 4299 4490 4287 4458 4458 28539048 0 0
28 2023-05-04 00:00:00+02:00 4149 4330 4123 4283 4283 15506868 0 0
29 2023-05-03 00:00:00+02:00 3975 4098 3968 4095 4095 14657028 0 0
30 2023-05-02 00:00:00+02:00 4037 4032 3917 3965 3965 11818133 0 0

View File

@@ -1,21 +1,19 @@
from .context import yfinance as yf
from .context import session_gbl
import unittest
import os
import datetime as _dt
import pytz as _tz
import numpy as _np
import pandas as _pd
import requests_cache
class TestPriceHistory(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -24,9 +22,7 @@ class TestPriceHistory(unittest.TestCase):
def test_daily_index(self):
tkrs = ["BHP.AX", "IMP.JO", "BP.L", "PNL.L", "INTC"]
intervals = ["1d", "1wk", "1mo"]
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
@@ -36,32 +32,55 @@ class TestPriceHistory(unittest.TestCase):
f = df.index.time == _dt.time(0)
self.assertTrue(f.all())
def test_download(self):
tkrs = ["BHP.AX", "IMP.JO", "BP.L", "PNL.L", "INTC"]
intervals = ["1d", "1wk", "1mo"]
for interval in intervals:
df = yf.download(tkrs, period="5y", interval=interval)
f = df.index.time == _dt.time(0)
self.assertTrue(f.all())
df_tkrs = df.columns.levels[1]
self.assertEqual(sorted(tkrs), sorted(df_tkrs))
def test_download_with_invalid_ticker(self):
#Checks if using an invalid symbol gives the same output as not using an invalid symbol in combination with a valid symbol (AAPL)
#Checks to make sure that invalid symbol handling for the date column is the same as the base case (no invalid symbols)
invalid_tkrs = ["AAPL", "ATVI"] #AAPL exists and ATVI does not exist
valid_tkrs = ["AAPL", "INTC"] #AAPL and INTC both exist
data_invalid_sym = yf.download(invalid_tkrs, start='2023-11-16', end='2023-11-17')
data_valid_sym = yf.download(valid_tkrs, start='2023-11-16', end='2023-11-17')
self.assertEqual(data_invalid_sym['Close']['AAPL']['2023-11-16'],data_valid_sym['Close']['AAPL']['2023-11-16'])
def test_duplicatingHourly(self):
tkrs = ["IMP.JO", "BHG.JO", "SSW.JO", "BP.L", "INTC"]
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
tz = dat._get_ticker_tz(debug_mode=False, proxy=None, timeout=None)
tz = dat._get_ticker_tz(proxy=None, timeout=None)
dt_utc = _tz.timezone("UTC").localize(_dt.datetime.utcnow())
dt = dt_utc.astimezone(_tz.timezone(tz))
df = dat.history(start=dt.date() - _dt.timedelta(days=1), interval="1h")
start_d = dt.date() - _dt.timedelta(days=7)
df = dat.history(start=start_d, interval="1h")
dt0 = df.index[-2]
dt1 = df.index[-1]
try:
self.assertNotEqual(dt0.hour, dt1.hour)
except:
except AssertionError:
print("Ticker = ", tkr)
raise
def test_duplicatingDaily(self):
tkrs = ["IMP.JO", "BHG.JO", "SSW.JO", "BP.L", "INTC"]
test_run = False
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
tz = dat._get_ticker_tz(debug_mode=False, proxy=None, timeout=None)
tz = dat._get_ticker_tz(proxy=None, timeout=None)
dt_utc = _tz.timezone("UTC").localize(_dt.datetime.utcnow())
dt = dt_utc.astimezone(_tz.timezone(tz))
@@ -75,7 +94,7 @@ class TestPriceHistory(unittest.TestCase):
dt1 = df.index[-1]
try:
self.assertNotEqual(dt0, dt1)
except:
except AssertionError:
print("Ticker = ", tkr)
raise
@@ -87,7 +106,7 @@ class TestPriceHistory(unittest.TestCase):
test_run = False
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
tz = dat._get_ticker_tz(debug_mode=False, proxy=None, timeout=None)
tz = dat._get_ticker_tz(proxy=None, timeout=None)
dt = _tz.timezone(tz).localize(_dt.datetime.now())
if dt.date().weekday() not in [1, 2, 3, 4]:
@@ -99,7 +118,7 @@ class TestPriceHistory(unittest.TestCase):
dt1 = df.index[-1]
try:
self.assertNotEqual(dt0.week, dt1.week)
except:
except AssertionError:
print("Ticker={}: Last two rows within same week:".format(tkr))
print(df.iloc[df.shape[0] - 2:])
raise
@@ -107,27 +126,121 @@ class TestPriceHistory(unittest.TestCase):
if not test_run:
self.skipTest("Skipping test_duplicatingWeekly() because not possible to fail Monday/weekend")
def test_pricesEventsMerge(self):
# Test case: dividend occurs after last row in price data
tkr = 'INTC'
start_d = _dt.date(2022, 1, 1)
end_d = _dt.date(2023, 1, 1)
df = yf.Ticker(tkr, session=self.session).history(interval='1d', start=start_d, end=end_d)
div = 1.0
future_div_dt = df.index[-1] + _dt.timedelta(days=1)
if future_div_dt.weekday() in [5, 6]:
future_div_dt += _dt.timedelta(days=1) * (7 - future_div_dt.weekday())
divs = _pd.DataFrame(data={"Dividends":[div]}, index=[future_div_dt])
df2 = yf.utils.safe_merge_dfs(df.drop(['Dividends', 'Stock Splits'], axis=1), divs, '1d')
self.assertIn(future_div_dt, df2.index)
self.assertIn("Dividends", df2.columns)
self.assertEqual(df2['Dividends'].iloc[-1], div)
def test_pricesEventsMerge_bug(self):
# Reproduce exception when merging intraday prices with future dividend
interval = '30m'
df_index = []
d = 13
for h in range(0, 16):
for m in [0, 30]:
df_index.append(_dt.datetime(2023, 9, d, h, m))
df_index.append(_dt.datetime(2023, 9, d, 16))
df = _pd.DataFrame(index=df_index)
df.index = _pd.to_datetime(df.index)
df['Close'] = 1.0
div = 1.0
future_div_dt = _dt.datetime(2023, 9, 14, 10)
divs = _pd.DataFrame(data={"Dividends":[div]}, index=[future_div_dt])
yf.utils.safe_merge_dfs(df, divs, interval)
# No exception = test pass
def test_intraDayWithEvents(self):
tkrs = ["BHP.AX", "IMP.JO", "BP.L", "PNL.L", "INTC"]
test_run = False
for tkr in tkrs:
start_d = _dt.date.today() - _dt.timedelta(days=59)
end_d = None
df_daily = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="1d", actions=True)
df_daily_divs = df_daily["Dividends"][df_daily["Dividends"] != 0]
if df_daily_divs.shape[0] == 0:
continue
last_div_date = df_daily_divs.index[-1]
start_d = last_div_date.date()
end_d = last_div_date.date() + _dt.timedelta(days=1)
df_intraday = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="15m", actions=True)
self.assertTrue((df_intraday["Dividends"] != 0.0).any())
df_intraday_divs = df_intraday["Dividends"][df_intraday["Dividends"] != 0]
df_intraday_divs.index = df_intraday_divs.index.floor('D')
self.assertTrue(df_daily_divs.equals(df_intraday_divs))
test_run = True
if not test_run:
self.skipTest("Skipping test_intraDayWithEvents() because no tickers had a dividend in last 60 days")
def test_intraDayWithEvents_tase(self):
# TASE dividend release pre-market, doesn't merge nicely with intra-day data so check still present
tkr = "ICL.TA"
# tkr = "ESLT.TA"
# tkr = "ONE.TA"
# tkr = "MGDL.TA"
start_d = _dt.date.today() - _dt.timedelta(days=60)
end_d = None
df_daily = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="1d", actions=True)
df_daily_divs = df_daily["Dividends"][df_daily["Dividends"] != 0]
if df_daily_divs.shape[0] == 0:
self.skipTest("Skipping test_intraDayWithEvents() because 'ICL.TA' has no dividend in last 60 days")
tase_tkrs = ["ICL.TA", "ESLT.TA", "ONE.TA", "MGDL.TA"]
test_run = False
for tkr in tase_tkrs:
start_d = _dt.date.today() - _dt.timedelta(days=59)
end_d = None
df_daily = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="1d", actions=True)
df_daily_divs = df_daily["Dividends"][df_daily["Dividends"] != 0]
if df_daily_divs.shape[0] == 0:
continue
last_div_date = df_daily_divs.index[-1]
start_d = last_div_date.date()
end_d = last_div_date.date() + _dt.timedelta(days=1)
df = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="15m", actions=True)
self.assertTrue((df["Dividends"] != 0.0).any())
last_div_date = df_daily_divs.index[-1]
start_d = last_div_date.date()
end_d = last_div_date.date() + _dt.timedelta(days=1)
df_intraday = yf.Ticker(tkr, session=self.session).history(start=start_d, end=end_d, interval="15m", actions=True)
self.assertTrue((df_intraday["Dividends"] != 0.0).any())
df_intraday_divs = df_intraday["Dividends"][df_intraday["Dividends"] != 0]
df_intraday_divs.index = df_intraday_divs.index.floor('D')
self.assertTrue(df_daily_divs.equals(df_intraday_divs))
test_run = True
if not test_run:
self.skipTest("Skipping test_intraDayWithEvents_tase() because no tickers had a dividend in last 60 days")
def test_dailyWithEvents(self):
start_d = _dt.date(2022, 1, 1)
end_d = _dt.date(2023, 1, 1)
tkr_div_dates = {'BHP.AX': [_dt.date(2022, 9, 1), _dt.date(2022, 2, 24)], # Yahoo claims 23-Feb but wrong because DST
'IMP.JO': [_dt.date(2022, 9, 21), _dt.date(2022, 3, 16)],
'BP.L': [_dt.date(2022, 11, 10), _dt.date(2022, 8, 11), _dt.date(2022, 5, 12),
_dt.date(2022, 2, 17)],
'INTC': [_dt.date(2022, 11, 4), _dt.date(2022, 8, 4), _dt.date(2022, 5, 5),
_dt.date(2022, 2, 4)]}
for tkr, dates in tkr_div_dates.items():
df = yf.Ticker(tkr, session=self.session).history(interval='1d', start=start_d, end=end_d)
df_divs = df[df['Dividends'] != 0].sort_index(ascending=False)
try:
self.assertTrue((df_divs.index.date == dates).all())
except AssertionError:
print(f'- ticker = {tkr}')
print('- response:')
print(df_divs.index.date)
print('- answer:')
print(dates)
raise
def test_dailyWithEvents_bugs(self):
# Reproduce issue #521
tkr1 = "QQQ"
tkr2 = "GDX"
@@ -139,7 +252,7 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df2["Dividends"] > 0) | (df2["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{} missing these dates: {}".format(tkr1, missing_from_df1))
@@ -154,13 +267,22 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df1["Dividends"] > 0) | (df1["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{}-with-events missing these dates: {}".format(tkr, missing_from_df1))
print("{}-without-events missing these dates: {}".format(tkr, missing_from_df2))
raise
# Reproduce issue #1634 - 1d dividend out-of-range, should be prepended to prices
div_dt = _pd.Timestamp(2022, 7, 21).tz_localize("America/New_York")
df_dividends = _pd.DataFrame(data={"Dividends":[1.0]}, index=[div_dt])
df_prices = _pd.DataFrame(data={c:[1.0] for c in yf.const._PRICE_COLNAMES_}|{'Volume':0}, index=[div_dt+_dt.timedelta(days=1)])
df_merged = yf.utils.safe_merge_dfs(df_prices, df_dividends, '1d')
self.assertEqual(df_merged.shape[0], 2)
self.assertTrue(df_merged[df_prices.columns].iloc[1:].equals(df_prices))
self.assertEqual(df_merged.index[0], div_dt)
def test_weeklyWithEvents(self):
# Reproduce issue #521
tkr1 = "QQQ"
@@ -173,7 +295,7 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df2["Dividends"] > 0) | (df2["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{} missing these dates: {}".format(tkr1, missing_from_df1))
@@ -188,7 +310,7 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df1["Dividends"] > 0) | (df1["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{}-with-events missing these dates: {}".format(tkr, missing_from_df1))
@@ -206,7 +328,7 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df2["Dividends"] > 0) | (df2["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{} missing these dates: {}".format(tkr1, missing_from_df1))
@@ -221,25 +343,40 @@ class TestPriceHistory(unittest.TestCase):
self.assertTrue(((df1["Dividends"] > 0) | (df1["Stock Splits"] > 0)).any())
try:
self.assertTrue(df1.index.equals(df2.index))
except:
except AssertionError:
missing_from_df1 = df2.index.difference(df1.index)
missing_from_df2 = df1.index.difference(df2.index)
print("{}-with-events missing these dates: {}".format(tkr, missing_from_df1))
print("{}-without-events missing these dates: {}".format(tkr, missing_from_df2))
raise
def test_monthlyWithEvents2(self):
# Simply check no exception from internal merge
dfm = yf.Ticker("ABBV").history(period="max", interval="1mo")
dfd = yf.Ticker("ABBV").history(period="max", interval="1d")
dfd = dfd[dfd.index > dfm.index[0]]
dfm_divs = dfm[dfm['Dividends'] != 0]
dfd_divs = dfd[dfd['Dividends'] != 0]
self.assertEqual(dfm_divs.shape[0], dfd_divs.shape[0])
dfm = yf.Ticker("F").history(period="50mo", interval="1mo")
dfd = yf.Ticker("F").history(period="50mo", interval="1d")
dfd = dfd[dfd.index > dfm.index[0]]
dfm_divs = dfm[dfm['Dividends'] != 0]
dfd_divs = dfd[dfd['Dividends'] != 0]
self.assertEqual(dfm_divs.shape[0], dfd_divs.shape[0])
def test_tz_dst_ambiguous(self):
# Reproduce issue #1100
try:
yf.Ticker("ESLT.TA", session=self.session).history(start="2002-10-06", end="2002-10-09", interval="1d")
except _tz.exceptions.AmbiguousTimeError:
raise Exception("Ambiguous DST issue not resolved")
def test_dst_fix(self):
# Daily intervals should start at time 00:00. But for some combinations of date and timezone,
# Daily intervals should start at time 00:00. But for some combinations of date and timezone,
# Yahoo has time off by few hours (e.g. Brazil 23:00 around Jan-2022). Suspect DST problem.
# The clue is (a) minutes=0 and (b) hour near 0.
# The clue is (a) minutes=0 and (b) hour near 0.
# Obviously Yahoo meant 00:00, so ensure this doesn't affect date conversion.
# The correction is successful if no days are weekend, and weekly data begins Monday
@@ -257,10 +394,50 @@ class TestPriceHistory(unittest.TestCase):
df = dat.history(start=start, end=end, interval=interval)
try:
self.assertTrue((df.index.weekday == 0).all())
except:
except AssertionError:
print("Weekly data not aligned to Monday")
raise
def test_prune_post_intraday_us(self):
# Half-day at USA Thanksgiving. Yahoo normally
# returns an interval starting when regular trading closes,
# even if prepost=False.
# Setup
tkr = "AMZN"
special_day = _dt.date(2023, 11, 24)
time_early_close = _dt.time(13)
dat = yf.Ticker(tkr, session=self.session)
# Run
start_d = special_day - _dt.timedelta(days=7)
end_d = special_day + _dt.timedelta(days=7)
df = dat.history(start=start_d, end=end_d, interval="1h", prepost=False, keepna=True)
tg_last_dt = df.loc[str(special_day)].index[-1]
self.assertTrue(tg_last_dt.time() < time_early_close)
# Test no other afternoons (or mornings) were pruned
start_d = _dt.date(special_day.year, 1, 1)
end_d = _dt.date(special_day.year+1, 1, 1)
df = dat.history(start=start_d, end=end_d, interval="1h", prepost=False, keepna=True)
last_dts = _pd.Series(df.index).groupby(df.index.date).last()
dfd = dat.history(start=start_d, end=end_d, interval='1d', prepost=False, keepna=True)
self.assertTrue(_np.equal(dfd.index.date, _pd.to_datetime(last_dts.index).date).all())
def test_prune_post_intraday_asx(self):
# Setup
tkr = "BHP.AX"
# No early closes in 2023
dat = yf.Ticker(tkr, session=self.session)
# Test no other afternoons (or mornings) were pruned
start_d = _dt.date(2023, 1, 1)
end_d = _dt.date(2023+1, 1, 1)
df = dat.history(start=start_d, end=end_d, interval="1h", prepost=False, keepna=True)
last_dts = _pd.Series(df.index).groupby(df.index.date).last()
dfd = dat.history(start=start_d, end=end_d, interval='1d', prepost=False, keepna=True)
self.assertTrue(_np.equal(dfd.index.date, _pd.to_datetime(last_dts.index).date).all())
def test_weekly_2rows_fix(self):
tkr = "AMZN"
start = _dt.date.today() - _dt.timedelta(days=14)
@@ -270,46 +447,100 @@ class TestPriceHistory(unittest.TestCase):
df = dat.history(start=start, interval="1wk")
self.assertTrue((df.index.weekday == 0).all())
def test_repair_100x_weekly(self):
def test_aggregate_capital_gains(self):
# Setup
tkr = "FXAIX"
dat = yf.Ticker(tkr, session=self.session)
start = "2017-12-31"
end = "2019-12-31"
interval = "3mo"
dat.history(start=start, end=end, interval=interval)
class TestPriceRepair(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
if cls.session is not None:
cls.session.close()
def test_types(self):
tkr = 'INTC'
dat = yf.Ticker(tkr, session=self.session)
data = dat.history(period="3mo", interval="1d", prepost=True, repair=True)
self.assertIsInstance(data, _pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
reconstructed = dat._lazy_load_price_history()._reconstruct_intervals_batch(data, "1wk", True)
self.assertIsInstance(reconstructed, _pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_reconstruct_2m(self):
# 2m repair requires 1m data.
# Yahoo restricts 1m fetches to 7 days max within last 30 days.
# Need to test that '_reconstruct_intervals_batch()' can handle this.
tkrs = ["BHP.AX", "IMP.JO", "BP.L", "PNL.L", "INTC"]
dt_now = _pd.Timestamp.utcnow()
td_60d = _dt.timedelta(days=60)
# Round time for 'requests_cache' reuse
dt_now = dt_now.ceil("1h")
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
end_dt = dt_now
start_dt = end_dt - td_60d
dat.history(start=start_dt, end=end_dt, interval="2m", repair=True)
def test_repair_100x_random_weekly(self):
# Setup:
tkr = "PNL.L"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.info["exchangeTimezoneName"]
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
data_cols = ["Low", "High", "Open", "Close", "Adj Close"]
df = _pd.DataFrame(data={"Open": [470.5, 473.5, 474.5, 470],
"High": [476, 476.5, 477, 480],
"Low": [470.5, 470, 465.5, 468.26],
"Close": [475, 473.5, 472, 473.5],
"Adj Close": [475, 473.5, 472, 473.5],
"Adj Close": [470.1, 468.6, 467.1, 468.6],
"Volume": [2295613, 2245604, 3000287, 2635611]},
index=_pd.to_datetime([_dt.date(2022, 10, 23),
_dt.date(2022, 10, 16),
_dt.date(2022, 10, 9),
_dt.date(2022, 10, 2)]))
index=_pd.to_datetime([_dt.date(2022, 10, 24),
_dt.date(2022, 10, 17),
_dt.date(2022, 10, 10),
_dt.date(2022, 10, 3)]))
df = df.sort_index()
df.index.name = "Date"
df_bad = df.copy()
df_bad.loc["2022-10-23", "Close"] *= 100
df_bad.loc["2022-10-16", "Low"] *= 100
df_bad.loc["2022-10-2", "Open"] *= 100
df_bad.loc["2022-10-24", "Close"] *= 100
df_bad.loc["2022-10-17", "Low"] *= 100
df_bad.loc["2022-10-03", "Open"] *= 100
df.index = df.index.tz_localize(tz_exchange)
df_bad.index = df_bad.index.tz_localize(tz_exchange)
# Run test
df_repaired = dat._fix_unit_mixups(df_bad, "1wk", tz_exchange)
df_repaired = hist._fix_unit_random_mixups(df_bad, "1wk", tz_exchange, prepost=False)
# First test - no errors left
for c in data_cols:
try:
self.assertTrue(_np.isclose(df_repaired[c], df[c], rtol=1e-2).all())
except:
except AssertionError:
print(df[c])
print(df_repaired[c])
raise
# Second test - all differences should be either ~1x or ~100x
ratio = df_bad[data_cols].values / df[data_cols].values
ratio = ratio.round(2)
@@ -321,24 +552,28 @@ class TestPriceHistory(unittest.TestCase):
f_1 = ratio == 1
self.assertTrue((f_100 | f_1).all())
def test_repair_100x_weekly_preSplit(self):
self.assertTrue("Repaired?" in df_repaired.columns)
self.assertFalse(df_repaired["Repaired?"].isna().any())
def test_repair_100x_random_weekly_preSplit(self):
# PNL.L has a stock-split in 2022. Sometimes requesting data before 2022 is not split-adjusted.
tkr = "PNL.L"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.info["exchangeTimezoneName"]
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
data_cols = ["Low", "High", "Open", "Close", "Adj Close"]
df = _pd.DataFrame(data={"Open": [400, 398, 392.5, 417],
"High": [421, 425, 419, 420.5],
"Low": [400, 380.5, 376.5, 396],
"Close": [410, 409.5, 402, 399],
"Adj Close": [398.02, 397.53, 390.25, 387.34],
df = _pd.DataFrame(data={"Open": [400, 398, 392.5, 417],
"High": [421, 425, 419, 420.5],
"Low": [400, 380.5, 376.5, 396],
"Close": [410, 409.5, 402, 399],
"Adj Close": [393.91, 393.43, 386.22, 383.34],
"Volume": [3232600, 3773900, 10835000, 4257900]},
index=_pd.to_datetime([_dt.date(2020, 3, 30),
_dt.date(2020, 3, 23),
_dt.date(2020, 3, 16),
_dt.date(2020, 3, 9)]))
index=_pd.to_datetime([_dt.date(2020, 3, 30),
_dt.date(2020, 3, 23),
_dt.date(2020, 3, 16),
_dt.date(2020, 3, 9)]))
df = df.sort_index()
# Simulate data missing split-adjustment:
df[data_cols] *= 100.0
@@ -353,13 +588,13 @@ class TestPriceHistory(unittest.TestCase):
df.index = df.index.tz_localize(tz_exchange)
df_bad.index = df_bad.index.tz_localize(tz_exchange)
df_repaired = dat._fix_unit_mixups(df_bad, "1wk", tz_exchange)
df_repaired = hist._fix_unit_random_mixups(df_bad, "1wk", tz_exchange, prepost=False)
# First test - no errors left
for c in data_cols:
try:
self.assertTrue(_np.isclose(df_repaired[c], df[c], rtol=1e-2).all())
except:
except AssertionError:
print("Mismatch in column", c)
print("- df_repaired:")
print(df_repaired[c])
@@ -378,10 +613,14 @@ class TestPriceHistory(unittest.TestCase):
f_1 = ratio == 1
self.assertTrue((f_100 | f_1).all())
def test_repair_100x_daily(self):
self.assertTrue("Repaired?" in df_repaired.columns)
self.assertFalse(df_repaired["Repaired?"].isna().any())
def test_repair_100x_random_daily(self):
tkr = "PNL.L"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.info["exchangeTimezoneName"]
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
data_cols = ["Low", "High", "Open", "Close", "Adj Close"]
df = _pd.DataFrame(data={"Open": [478, 476, 476, 472],
@@ -390,10 +629,10 @@ class TestPriceHistory(unittest.TestCase):
"Close": [475.5, 475.5, 474.5, 475],
"Adj Close": [475.5, 475.5, 474.5, 475],
"Volume": [436414, 485947, 358067, 287620]},
index=_pd.to_datetime([_dt.date(2022, 11, 1),
_dt.date(2022, 10, 31),
_dt.date(2022, 10, 28),
_dt.date(2022, 10, 27)]))
index=_pd.to_datetime([_dt.date(2022, 11, 1),
_dt.date(2022, 10, 31),
_dt.date(2022, 10, 28),
_dt.date(2022, 10, 27)]))
df = df.sort_index()
df.index.name = "Date"
df_bad = df.copy()
@@ -403,7 +642,7 @@ class TestPriceHistory(unittest.TestCase):
df.index = df.index.tz_localize(tz_exchange)
df_bad.index = df_bad.index.tz_localize(tz_exchange)
df_repaired = dat._fix_unit_mixups(df_bad, "1d", tz_exchange)
df_repaired = hist._fix_unit_random_mixups(df_bad, "1d", tz_exchange, prepost=False)
# First test - no errors left
for c in data_cols:
@@ -420,10 +659,69 @@ class TestPriceHistory(unittest.TestCase):
f_1 = ratio == 1
self.assertTrue((f_100 | f_1).all())
self.assertTrue("Repaired?" in df_repaired.columns)
self.assertFalse(df_repaired["Repaired?"].isna().any())
def test_repair_100x_block_daily(self):
# Some 100x errors are not sporadic.
# Sometimes Yahoo suddenly shifts from cents->$ from some recent date.
tkrs = ['AET.L', 'SSW.JO']
for tkr in tkrs:
for interval in ['1d', '1wk']:
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
data_cols = ["Low", "High", "Open", "Close", "Adj Close"]
_dp = os.path.dirname(__file__)
fp = os.path.join(_dp, "data", tkr.replace('.','-') + '-' + interval + "-100x-error.csv")
if not os.path.isfile(fp):
continue
df_bad = _pd.read_csv(fp, index_col="Date")
df_bad.index = _pd.to_datetime(df_bad.index, utc=True).tz_convert(tz_exchange)
df_bad = df_bad.sort_index()
df = df_bad.copy()
fp = os.path.join(_dp, "data", tkr.replace('.','-') + '-' + interval + "-100x-error-fixed.csv")
df = _pd.read_csv(fp, index_col="Date")
df.index = _pd.to_datetime(df.index, utc=True).tz_convert(tz_exchange)
df = df.sort_index()
df_repaired = hist._fix_unit_switch(df_bad, interval, tz_exchange)
df_repaired = df_repaired.sort_index()
# First test - no errors left
for c in data_cols:
try:
self.assertTrue(_np.isclose(df_repaired[c], df[c], rtol=1e-2).all())
except:
print("- repaired:")
print(df_repaired[c])
print("- correct:")
print(df[c])
print(f"TEST FAIL on column '{c}' (tkr={tkr} interval={interval})")
raise
# Second test - all differences should be either ~1x or ~100x
ratio = df_bad[data_cols].values / df[data_cols].values
ratio = ratio.round(2)
# - round near-100 ratio to 100:
f = ratio > 90
ratio[f] = (ratio[f] / 10).round().astype(int) * 10 # round ratio to nearest 10
# - now test
f_100 = (ratio == 100) | (ratio == 0.01)
f_1 = ratio == 1
self.assertTrue((f_100 | f_1).all())
self.assertTrue("Repaired?" in df_repaired.columns)
self.assertFalse(df_repaired["Repaired?"].isna().any())
def test_repair_zeroes_daily(self):
tkr = "BBIL.L"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.info["exchangeTimezoneName"]
hist = dat._lazy_load_price_history()
tz_exchange = dat.fast_info["timezone"]
df_bad = _pd.DataFrame(data={"Open": [0, 102.04, 102.04],
"High": [0, 102.1, 102.11],
@@ -431,14 +729,14 @@ class TestPriceHistory(unittest.TestCase):
"Close": [103.03, 102.05, 102.08],
"Adj Close": [102.03, 102.05, 102.08],
"Volume": [560, 137, 117]},
index=_pd.to_datetime([_dt.datetime(2022, 11, 1),
_dt.datetime(2022, 10, 31),
_dt.datetime(2022, 10, 30)]))
index=_pd.to_datetime([_dt.datetime(2022, 11, 1),
_dt.datetime(2022, 10, 31),
_dt.datetime(2022, 10, 30)]))
df_bad = df_bad.sort_index()
df_bad.index.name = "Date"
df_bad.index = df_bad.index.tz_localize(tz_exchange)
repaired_df = dat._fix_zeroes(df_bad, "1d", tz_exchange)
repaired_df = hist._fix_zeroes(df_bad, "1d", tz_exchange, prepost=False)
correct_df = df_bad.copy()
correct_df.loc["2022-11-01", "Open"] = 102.080002
@@ -447,53 +745,209 @@ class TestPriceHistory(unittest.TestCase):
for c in ["Open", "Low", "High", "Close"]:
self.assertTrue(_np.isclose(repaired_df[c], correct_df[c], rtol=1e-8).all())
self.assertTrue("Repaired?" in repaired_df.columns)
self.assertFalse(repaired_df["Repaired?"].isna().any())
def test_repair_zeroes_daily_adjClose(self):
# Test that 'Adj Close' is reconstructed correctly,
# particularly when a dividend occurred within 1 day.
tkr = "INTC"
df = _pd.DataFrame(data={"Open": [28.95, 28.65, 29.55, 29.62, 29.25],
"High": [29.12, 29.27, 29.65, 31.17, 30.30],
"Low": [28.21, 28.43, 28.61, 29.53, 28.80],
"Close": [28.24, 29.05, 28.69, 30.32, 30.19],
"Adj Close": [28.12, 28.93, 28.57, 29.83, 29.70],
"Volume": [36e6, 51e6, 49e6, 58e6, 62e6],
"Dividends": [0, 0, 0.365, 0, 0]},
index=_pd.to_datetime([_dt.datetime(2023, 2, 8),
_dt.datetime(2023, 2, 7),
_dt.datetime(2023, 2, 6),
_dt.datetime(2023, 2, 3),
_dt.datetime(2023, 2, 2)]))
df = df.sort_index()
df.index.name = "Date"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
df.index = df.index.tz_localize(tz_exchange)
hist = dat._lazy_load_price_history()
rtol = 5e-3
for i in [0, 1, 2]:
df_slice = df.iloc[i:i+3]
for j in range(3):
df_slice_bad = df_slice.copy()
df_slice_bad.loc[df_slice_bad.index[j], "Adj Close"] = 0.0
df_slice_bad_repaired = hist._fix_zeroes(df_slice_bad, "1d", tz_exchange, prepost=False)
for c in ["Close", "Adj Close"]:
self.assertTrue(_np.isclose(df_slice_bad_repaired[c], df_slice[c], rtol=rtol).all())
self.assertTrue("Repaired?" in df_slice_bad_repaired.columns)
self.assertFalse(df_slice_bad_repaired["Repaired?"].isna().any())
def test_repair_zeroes_hourly(self):
tkr = "INTC"
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.info["exchangeTimezoneName"]
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
df_bad = _pd.DataFrame(data={"Open": [29.68, 29.49, 29.545, _np.nan, 29.485],
"High": [29.68, 29.625, 29.58, _np.nan, 29.49],
"Low": [29.46, 29.4, 29.45, _np.nan, 29.31],
"Close": [29.485, 29.545, 29.485, _np.nan, 29.325],
"Adj Close": [29.485, 29.545, 29.485, _np.nan, 29.325],
"Volume": [3258528, 2140195, 1621010, 0, 0]},
index=_pd.to_datetime([_dt.datetime(2022,11,25, 9,30),
_dt.datetime(2022,11,25, 10,30),
_dt.datetime(2022,11,25, 11,30),
_dt.datetime(2022,11,25, 12,30),
_dt.datetime(2022,11,25, 13,00)]))
df_bad = df_bad.sort_index()
df_bad.index.name = "Date"
df_bad.index = df_bad.index.tz_localize(tz_exchange)
correct_df = hist.history(period="1wk", interval="1h", auto_adjust=False, repair=True)
repaired_df = dat._fix_zeroes(df_bad, "1h", tz_exchange)
df_bad = correct_df.copy()
bad_idx = correct_df.index[10]
df_bad.loc[bad_idx, "Open"] = _np.nan
df_bad.loc[bad_idx, "High"] = _np.nan
df_bad.loc[bad_idx, "Low"] = _np.nan
df_bad.loc[bad_idx, "Close"] = _np.nan
df_bad.loc[bad_idx, "Adj Close"] = _np.nan
df_bad.loc[bad_idx, "Volume"] = 0
repaired_df = hist._fix_zeroes(df_bad, "1h", tz_exchange, prepost=False)
correct_df = df_bad.copy()
idx = _pd.Timestamp(2022,11,25, 12,30).tz_localize(tz_exchange)
correct_df.loc[idx, "Open"] = 29.485001
correct_df.loc[idx, "High"] = 29.49
correct_df.loc[idx, "Low"] = 29.43
correct_df.loc[idx, "Close"] = 29.455
correct_df.loc[idx, "Adj Close"] = 29.455
correct_df.loc[idx, "Volume"] = 609164
for c in ["Open", "Low", "High", "Close"]:
try:
self.assertTrue(_np.isclose(repaired_df[c], correct_df[c], rtol=1e-7).all())
except:
except AssertionError:
print("COLUMN", c)
print("- repaired_df")
print(repaired_df)
print("- correct_df[c]:")
print(correct_df[c])
print("- diff:")
print(repaired_df[c] - correct_df[c])
raise
self.assertTrue("Repaired?" in repaired_df.columns)
self.assertFalse(repaired_df["Repaired?"].isna().any())
def test_repair_bad_stock_split(self):
# Stocks that split in 2022 but no problems in Yahoo data,
# so repair should change nothing
good_tkrs = ['AMZN', 'DXCM', 'FTNT', 'GOOG', 'GME', 'PANW', 'SHOP', 'TSLA']
good_tkrs += ['AEI', 'GHI', 'IRON', 'LXU', 'NUZE', 'RSLS', 'TISI']
good_tkrs += ['BOL.ST', 'TUI1.DE']
intervals = ['1d', '1wk', '1mo', '3mo']
for tkr in good_tkrs:
for interval in intervals:
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
_dp = os.path.dirname(__file__)
df_good = dat.history(start='2020-01-01', end=_dt.date.today(), interval=interval, auto_adjust=False)
repaired_df = hist._fix_bad_stock_split(df_good, interval, tz_exchange)
# Expect no change from repair
df_good = df_good.sort_index()
repaired_df = repaired_df.sort_index()
for c in ["Open", "Low", "High", "Close", "Adj Close", "Volume"]:
try:
self.assertTrue((repaired_df[c].to_numpy() == df_good[c].to_numpy()).all())
except:
print(f"tkr={tkr} interval={interval} COLUMN={c}")
df_dbg = df_good[[c]].join(repaired_df[[c]], lsuffix='.good', rsuffix='.repaired')
f_diff = repaired_df[c].to_numpy() != df_good[c].to_numpy()
print(df_dbg[f_diff | _np.roll(f_diff, 1) | _np.roll(f_diff, -1)])
raise
bad_tkrs = ['4063.T', 'ALPHA.PA', 'AV.L', 'CNE.L', 'MOB.ST', 'SPM.MI']
bad_tkrs.append('LA.V') # special case - stock split error is 3 years ago! why not fixed?
for tkr in bad_tkrs:
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
_dp = os.path.dirname(__file__)
interval = '1d'
fp = os.path.join(_dp, "data", tkr.replace('.','-')+'-'+interval+"-bad-stock-split.csv")
if not os.path.isfile(fp):
interval = '1wk'
fp = os.path.join(_dp, "data", tkr.replace('.','-')+'-'+interval+"-bad-stock-split.csv")
df_bad = _pd.read_csv(fp, index_col="Date")
df_bad.index = _pd.to_datetime(df_bad.index, utc=True)
repaired_df = hist._fix_bad_stock_split(df_bad, "1d", tz_exchange)
fp = os.path.join(_dp, "data", tkr.replace('.','-')+'-'+interval+"-bad-stock-split-fixed.csv")
correct_df = _pd.read_csv(fp, index_col="Date")
correct_df.index = _pd.to_datetime(correct_df.index)
repaired_df = repaired_df.sort_index()
correct_df = correct_df.sort_index()
for c in ["Open", "Low", "High", "Close", "Adj Close", "Volume"]:
try:
self.assertTrue(_np.isclose(repaired_df[c], correct_df[c], rtol=5e-6).all())
except AssertionError:
print(f"tkr={tkr} COLUMN={c}")
# print("- repaired_df")
# print(repaired_df)
# print("- correct_df[c]:")
# print(correct_df[c])
# print("- diff:")
# print(repaired_df[c] - correct_df[c])
raise
# Had very high price volatility in Jan-2021 around split date that could
# be mistaken for missing stock split adjustment. And old logic did think
# column 'High' required fixing - wrong!
sketchy_tkrs = ['FIZZ']
intervals = ['1wk']
for tkr in sketchy_tkrs:
for interval in intervals:
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
_dp = os.path.dirname(__file__)
df_good = hist.history(start='2020-11-30', end='2021-04-01', interval=interval, auto_adjust=False)
repaired_df = hist._fix_bad_stock_split(df_good, interval, tz_exchange)
# Expect no change from repair
df_good = df_good.sort_index()
repaired_df = repaired_df.sort_index()
for c in ["Open", "Low", "High", "Close", "Adj Close", "Volume"]:
try:
self.assertTrue((repaired_df[c].to_numpy() == df_good[c].to_numpy()).all())
except AssertionError:
print(f"tkr={tkr} interval={interval} COLUMN={c}")
df_dbg = df_good[[c]].join(repaired_df[[c]], lsuffix='.good', rsuffix='.repaired')
f_diff = repaired_df[c].to_numpy() != df_good[c].to_numpy()
print(df_dbg[f_diff | _np.roll(f_diff, 1) | _np.roll(f_diff, -1)])
raise
def test_repair_missing_div_adjust(self):
tkr = '8TRA.DE'
dat = yf.Ticker(tkr, session=self.session)
tz_exchange = dat.fast_info["timezone"]
hist = dat._lazy_load_price_history()
_dp = os.path.dirname(__file__)
df_bad = _pd.read_csv(os.path.join(_dp, "data", tkr.replace('.','-')+"-1d-missing-div-adjust.csv"), index_col="Date")
df_bad.index = _pd.to_datetime(df_bad.index)
repaired_df = hist._fix_missing_div_adjust(df_bad, "1d", tz_exchange)
correct_df = _pd.read_csv(os.path.join(_dp, "data", tkr.replace('.','-')+"-1d-missing-div-adjust-fixed.csv"), index_col="Date")
correct_df.index = _pd.to_datetime(correct_df.index)
repaired_df = repaired_df.sort_index()
correct_df = correct_df.sort_index()
for c in ["Open", "Low", "High", "Close", "Adj Close", "Volume"]:
try:
self.assertTrue(_np.isclose(repaired_df[c], correct_df[c], rtol=5e-6).all())
except:
print(f"tkr={tkr} COLUMN={c}")
print("- repaired_df")
print(repaired_df)
print("- correct_df[c]:")
print(correct_df[c])
print("- diff:")
print(repaired_df[c] - correct_df[c])
raise
if __name__ == '__main__':
unittest.main()
# # Run tests sequentially:
# import inspect
# test_src = inspect.getsource(TestPriceHistory)
# unittest.TestLoader.sortTestMethodsUsing = lambda _, x, y: (
# test_src.index(f"def {x}") - test_src.index(f"def {y}")
# )
# unittest.main(verbosity=2)

View File

@@ -9,28 +9,73 @@ Specific test class:
"""
import pandas as pd
import numpy as np
from .context import yfinance as yf
from .context import session_gbl
from yfinance.exceptions import YFNotImplementedError
import unittest
import requests_cache
from typing import Union, Any, get_args, _GenericAlias
from urllib.parse import urlparse, parse_qs, urlencode, urlunparse
# Set this to see the exact requests that are made during tests
DEBUG_LOG_REQUESTS = False
if DEBUG_LOG_REQUESTS:
import logging
logging.basicConfig(level=logging.DEBUG)
ticker_attributes = (
("major_holders", pd.DataFrame),
("institutional_holders", pd.DataFrame),
("mutualfund_holders", pd.DataFrame),
("insider_transactions", pd.DataFrame),
("insider_purchases", pd.DataFrame),
("insider_roster_holders", pd.DataFrame),
("splits", pd.Series),
("actions", pd.DataFrame),
("shares", pd.DataFrame),
("info", dict),
("calendar", dict),
("recommendations", Union[pd.DataFrame, dict]),
("recommendations_summary", Union[pd.DataFrame, dict]),
("upgrades_downgrades", Union[pd.DataFrame, dict]),
("earnings", pd.DataFrame),
("quarterly_earnings", pd.DataFrame),
("quarterly_cashflow", pd.DataFrame),
("cashflow", pd.DataFrame),
("quarterly_balance_sheet", pd.DataFrame),
("balance_sheet", pd.DataFrame),
("quarterly_income_stmt", pd.DataFrame),
("income_stmt", pd.DataFrame),
("analyst_price_target", pd.DataFrame),
("revenue_forecasts", pd.DataFrame),
("sustainability", pd.DataFrame),
("options", tuple),
("news", Any),
("earnings_trend", pd.DataFrame),
("earnings_dates", pd.DataFrame),
("earnings_forecasts", pd.DataFrame),
)
def assert_attribute_type(testClass: unittest.TestCase, instance, attribute_name, expected_type):
try:
attribute = getattr(instance, attribute_name)
if attribute is not None and expected_type is not Any:
err_msg = f'{attribute_name} type is {type(attribute)} not {expected_type}'
if isinstance(expected_type, _GenericAlias) and expected_type.__origin__ is Union:
allowed_types = get_args(expected_type)
testClass.assertTrue(isinstance(attribute, allowed_types), err_msg)
else:
testClass.assertEqual(type(attribute), expected_type, err_msg)
except Exception:
testClass.assertRaises(
YFNotImplementedError, lambda: getattr(instance, attribute_name)
)
class TestTicker(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
cls.proxy = None
@classmethod
def tearDownClass(cls):
@@ -41,101 +86,90 @@ class TestTicker(unittest.TestCase):
tkrs = ["IMP.JO", "BHG.JO", "SSW.JO", "BP.L", "INTC"]
for tkr in tkrs:
# First step: remove ticker from tz-cache
yf.utils.get_tz_cache().store(tkr, None)
yf.cache.get_tz_cache().store(tkr, None)
# Test:
dat = yf.Ticker(tkr, session=self.session)
tz = dat._get_ticker_tz(debug_mode=False, proxy=None, timeout=None)
tz = dat._get_ticker_tz(proxy=None, timeout=5)
self.assertIsNotNone(tz)
def test_badTicker(self):
# Check yfinance doesn't die when ticker delisted
tkr = "AM2Z.TA"
tkr = "DJI" # typo of "^DJI"
dat = yf.Ticker(tkr, session=self.session)
dat.history(period="1wk")
dat.history(start="2022-01-01")
dat.history(start="2022-01-01", end="2022-03-01")
yf.download([tkr], period="1wk")
dat.isin
dat.major_holders
dat.institutional_holders
dat.mutualfund_holders
dat.dividends
dat.splits
dat.actions
dat.shares
dat.get_shares_full()
dat.info
dat.calendar
dat.recommendations
dat.earnings
dat.quarterly_earnings
dat.income_stmt
dat.quarterly_income_stmt
dat.balance_sheet
dat.quarterly_balance_sheet
dat.cashflow
dat.quarterly_cashflow
dat.recommendations_summary
dat.analyst_price_target
dat.revenue_forecasts
dat.sustainability
dat.options
dat.news
dat.earnings_trend
dat.earnings_dates
dat.earnings_forecasts
yf.download([tkr], period="1wk", threads=False, ignore_tz=False)
yf.download([tkr], period="1wk", threads=True, ignore_tz=False)
yf.download([tkr], period="1wk", threads=False, ignore_tz=True)
yf.download([tkr], period="1wk", threads=True, ignore_tz=True)
for k in dat.fast_info:
dat.fast_info[k]
for attribute_name, attribute_type in ticker_attributes:
assert_attribute_type(self, dat, attribute_name, attribute_type)
with self.assertRaises(YFNotImplementedError):
assert isinstance(dat.earnings, pd.Series)
assert dat.earnings.empty
assert isinstance(dat.dividends, pd.Series)
assert dat.dividends.empty
assert isinstance(dat.splits, pd.Series)
assert dat.splits.empty
assert isinstance(dat.capital_gains, pd.Series)
assert dat.capital_gains.empty
with self.assertRaises(YFNotImplementedError):
assert isinstance(dat.shares, pd.DataFrame)
assert dat.shares.empty
assert isinstance(dat.actions, pd.DataFrame)
assert dat.actions.empty
def test_goodTicker(self):
# that yfinance works when full api is called on same instance of ticker
tkrs = ["IBM"]
tkrs.append("QCSTIX") # weird ticker, no price history but has previous close
for tkr in tkrs:
dat = yf.Ticker(tkr, session=self.session)
dat.history(period="1wk")
dat.history(start="2022-01-01")
dat.history(start="2022-01-01", end="2022-03-01")
yf.download([tkr], period="1wk", threads=False, ignore_tz=False)
yf.download([tkr], period="1wk", threads=True, ignore_tz=False)
yf.download([tkr], period="1wk", threads=False, ignore_tz=True)
yf.download([tkr], period="1wk", threads=True, ignore_tz=True)
for k in dat.fast_info:
dat.fast_info[k]
for attribute_name, attribute_type in ticker_attributes:
assert_attribute_type(self, dat, attribute_name, attribute_type)
def test_goodTicker_withProxy(self):
tkr = "IBM"
dat = yf.Ticker(tkr, session=self.session)
dat.isin
dat.major_holders
dat.institutional_holders
dat.mutualfund_holders
dat.dividends
dat.splits
dat.actions
dat.shares
dat.get_shares_full()
dat.info
dat.calendar
dat.recommendations
dat.earnings
dat.quarterly_earnings
dat.income_stmt
dat.quarterly_income_stmt
dat.balance_sheet
dat.quarterly_balance_sheet
dat.cashflow
dat.quarterly_cashflow
dat.recommendations_summary
dat.analyst_price_target
dat.revenue_forecasts
dat.sustainability
dat.options
dat.news
dat.earnings_trend
dat.earnings_dates
dat.earnings_forecasts
dat = yf.Ticker(tkr, session=self.session, proxy=self.proxy)
dat._fetch_ticker_tz(proxy=None, timeout=5)
dat._get_ticker_tz(proxy=None, timeout=5)
dat.history(period="1wk")
dat.history(start="2022-01-01")
dat.history(start="2022-01-01", end="2022-03-01")
yf.download([tkr], period="1wk")
for attribute_name, attribute_type in ticker_attributes:
assert_attribute_type(self, dat, attribute_name, attribute_type)
class TestTickerHistory(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -144,35 +178,60 @@ class TestTickerHistory(unittest.TestCase):
def setUp(self):
# use a ticker that has dividends
self.ticker = yf.Ticker("IBM", session=self.session)
self.symbol = "IBM"
self.ticker = yf.Ticker(self.symbol, session=self.session)
self.symbols = ["AMZN", "MSFT", "NVDA"]
def tearDown(self):
self.ticker = None
def test_history(self):
with self.assertRaises(RuntimeError):
self.ticker.history_metadata
md = self.ticker.history_metadata
self.assertIn("IBM", md.values(), "metadata missing")
data = self.ticker.history("1y")
self.assertIn("IBM", self.ticker.history_metadata.values(), "metadata missing")
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_download(self):
for t in [False, True]:
for i in [False, True]:
data = yf.download(self.symbols, threads=t, ignore_tz=i)
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_no_expensive_calls_introduced(self):
"""
Make sure calling history to get price data has not introduced more calls to yahoo than absolutely necessary.
As doing other type of scraping calls than "query2.finance.yahoo.com/v8/finance/chart" to yahoo website
will quickly trigger spam-block when doing bulk download of history data.
"""
session = requests_cache.CachedSession(backend='memory')
ticker = yf.Ticker("GOOGL", session=session)
ticker.history("1y")
actual_urls_called = tuple([r.url for r in session.cache.filter()])
session.close()
expected_urls = (
'https://query2.finance.yahoo.com/v8/finance/chart/GOOGL?range=1y&interval=1d&includePrePost=False&events=div%2Csplits%2CcapitalGains',
)
self.assertEqual(expected_urls, actual_urls_called, "Different than expected url used to fetch history.")
symbol = "GOOGL"
period = "1y"
with requests_cache.CachedSession(backend="memory") as session:
ticker = yf.Ticker(symbol, session=session)
ticker.history(period=period)
actual_urls_called = [r.url for r in session.cache.filter()]
# Remove 'crumb' argument
for i in range(len(actual_urls_called)):
u = actual_urls_called[i]
parsed_url = urlparse(u)
query_params = parse_qs(parsed_url.query)
query_params.pop('crumb', None)
query_params.pop('cookie', None)
u = urlunparse(parsed_url._replace(query=urlencode(query_params, doseq=True)))
actual_urls_called[i] = u
actual_urls_called = tuple(actual_urls_called)
expected_urls = (
f"https://query2.finance.yahoo.com/v8/finance/chart/{symbol}?events=div%2Csplits%2CcapitalGains&includePrePost=False&interval=1d&range={period}",
)
self.assertEqual(
expected_urls,
actual_urls_called,
"Different than expected url used to fetch history."
)
def test_dividends(self):
data = self.ticker.dividends
self.assertIsInstance(data, pd.Series, "data has wrong type")
@@ -194,7 +253,7 @@ class TestTickerEarnings(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -207,46 +266,11 @@ class TestTickerEarnings(unittest.TestCase):
def tearDown(self):
self.ticker = None
def test_earnings(self):
data = self.ticker.earnings
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.earnings
self.assertIs(data, data_cached, "data not cached")
def test_quarterly_earnings(self):
data = self.ticker.quarterly_earnings
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.quarterly_earnings
self.assertIs(data, data_cached, "data not cached")
def test_earnings_forecasts(self):
data = self.ticker.earnings_forecasts
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.earnings_forecasts
self.assertIs(data, data_cached, "data not cached")
def test_earnings_dates(self):
data = self.ticker.earnings_dates
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.earnings_dates
self.assertIs(data, data_cached, "data not cached")
def test_earnings_trend(self):
data = self.ticker.earnings_trend
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.earnings_trend
self.assertIs(data, data_cached, "data not cached")
def test_earnings_dates_with_limit(self):
# use ticker with lots of historic earnings
ticker = yf.Ticker("IBM")
@@ -259,13 +283,50 @@ class TestTickerEarnings(unittest.TestCase):
data_cached = ticker.get_earnings_dates(limit=limit)
self.assertIs(data, data_cached, "data not cached")
# Below will fail because not ported to Yahoo API
# def test_earnings(self):
# data = self.ticker.earnings
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.earnings
# self.assertIs(data, data_cached, "data not cached")
# def test_quarterly_earnings(self):
# data = self.ticker.quarterly_earnings
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.quarterly_earnings
# self.assertIs(data, data_cached, "data not cached")
# def test_earnings_forecasts(self):
# data = self.ticker.earnings_forecasts
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.earnings_forecasts
# self.assertIs(data, data_cached, "data not cached")
# data_cached = self.ticker.earnings_dates
# self.assertIs(data, data_cached, "data not cached")
# def test_earnings_trend(self):
# data = self.ticker.earnings_trend
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.earnings_trend
# self.assertIs(data, data_cached, "data not cached")
class TestTickerHolders(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -302,13 +363,37 @@ class TestTickerHolders(unittest.TestCase):
data_cached = self.ticker.mutualfund_holders
self.assertIs(data, data_cached, "data not cached")
def test_insider_transactions(self):
data = self.ticker.insider_transactions
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.insider_transactions
self.assertIs(data, data_cached, "data not cached")
def test_insider_purchases(self):
data = self.ticker.insider_purchases
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.insider_purchases
self.assertIs(data, data_cached, "data not cached")
def test_insider_roster_holders(self):
data = self.ticker.insider_roster_holders
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.insider_roster_holders
self.assertIs(data, data_cached, "data not cached")
class TestTickerMiscFinancials(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -326,6 +411,24 @@ class TestTickerMiscFinancials(unittest.TestCase):
def tearDown(self):
self.ticker = None
def test_isin(self):
data = self.ticker.isin
self.assertIsInstance(data, str, "data has wrong type")
self.assertEqual("ARDEUT116159", data, "data is empty")
data_cached = self.ticker.isin
self.assertIs(data, data_cached, "data not cached")
def test_options(self):
data = self.ticker.options
self.assertIsInstance(data, tuple, "data has wrong type")
self.assertTrue(len(data) > 1, "data is empty")
def test_shares_full(self):
data = self.ticker.get_shares_full()
self.assertIsInstance(data, pd.Series, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_income_statement(self):
expected_keys = ["Total Revenue", "Basic EPS"]
expected_periods_days = 365
@@ -355,7 +458,6 @@ class TestTickerMiscFinancials(unittest.TestCase):
data = self.ticker.get_income_stmt(as_dict=True)
self.assertIsInstance(data, dict, "data has wrong type")
def test_quarterly_income_statement(self):
expected_keys = ["Total Revenue", "Basic EPS"]
expected_periods_days = 365//4
@@ -385,16 +487,6 @@ class TestTickerMiscFinancials(unittest.TestCase):
data = self.ticker.get_income_stmt(as_dict=True)
self.assertIsInstance(data, dict, "data has wrong type")
def test_quarterly_income_statement_old_fmt(self):
expected_row = "TotalRevenue"
data = self.ticker_old_fmt.get_income_stmt(freq="quarterly", legacy=True)
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
self.assertIn(expected_row, data.index, "Did not find expected row in index")
data_cached = self.ticker_old_fmt.get_income_stmt(freq="quarterly", legacy=True)
self.assertIs(data, data_cached, "data not cached")
def test_balance_sheet(self):
expected_keys = ["Total Assets", "Net PPE"]
expected_periods_days = 365
@@ -453,16 +545,6 @@ class TestTickerMiscFinancials(unittest.TestCase):
data = self.ticker.get_balance_sheet(as_dict=True, freq="quarterly")
self.assertIsInstance(data, dict, "data has wrong type")
def test_quarterly_balance_sheet_old_fmt(self):
expected_row = "TotalAssets"
data = self.ticker_old_fmt.get_balance_sheet(freq="quarterly", legacy=True)
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
self.assertIn(expected_row, data.index, "Did not find expected row in index")
data_cached = self.ticker_old_fmt.get_balance_sheet(freq="quarterly", legacy=True)
self.assertIs(data, data_cached, "data not cached")
def test_cash_flow(self):
expected_keys = ["Operating Cash Flow", "Net PPE Purchase And Sale"]
expected_periods_days = 365
@@ -521,16 +603,6 @@ class TestTickerMiscFinancials(unittest.TestCase):
data = self.ticker.get_cashflow(as_dict=True)
self.assertIsInstance(data, dict, "data has wrong type")
def test_quarterly_cashflow_old_fmt(self):
expected_row = "NetIncome"
data = self.ticker_old_fmt.get_cashflow(legacy=True, freq="quarterly")
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
self.assertIn(expected_row, data.index, "Did not find expected row in index")
data_cached = self.ticker_old_fmt.get_cashflow(legacy=True, freq="quarterly")
self.assertIs(data, data_cached, "data not cached")
def test_income_alt_names(self):
i1 = self.ticker.income_stmt
i2 = self.ticker.incomestmt
@@ -590,87 +662,49 @@ class TestTickerMiscFinancials(unittest.TestCase):
i2 = self.ticker.get_cashflow(freq="quarterly")
self.assertTrue(i1.equals(i2))
def test_sustainability(self):
data = self.ticker.sustainability
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.sustainability
self.assertIs(data, data_cached, "data not cached")
def test_recommendations(self):
data = self.ticker.recommendations
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.recommendations
self.assertIs(data, data_cached, "data not cached")
def test_recommendations_summary(self):
data = self.ticker.recommendations_summary
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.recommendations_summary
self.assertIs(data, data_cached, "data not cached")
def test_analyst_price_target(self):
data = self.ticker.analyst_price_target
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.analyst_price_target
self.assertIs(data, data_cached, "data not cached")
def test_revenue_forecasts(self):
data = self.ticker.revenue_forecasts
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.revenue_forecasts
self.assertIs(data, data_cached, "data not cached")
def test_calendar(self):
data = self.ticker.calendar
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.calendar
self.assertIs(data, data_cached, "data not cached")
def test_isin(self):
data = self.ticker.isin
self.assertIsInstance(data, str, "data has wrong type")
self.assertEqual("ARDEUT116159", data, "data is empty")
data_cached = self.ticker.isin
self.assertIs(data, data_cached, "data not cached")
def test_options(self):
data = self.ticker.options
self.assertIsInstance(data, tuple, "data has wrong type")
self.assertTrue(len(data) > 1, "data is empty")
def test_shares(self):
data = self.ticker.shares
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_shares_full(self):
data = self.ticker.get_shares_full()
self.assertIsInstance(data, pd.Series, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
def test_bad_freq_value_raises_exception(self):
self.assertRaises(ValueError, lambda: self.ticker.get_cashflow(freq="badarg"))
def test_calendar(self):
data = self.ticker.calendar
self.assertIsInstance(data, dict, "data has wrong type")
self.assertTrue(len(data) > 0, "data is empty")
self.assertIn("Earnings Date", data.keys(), "data missing expected key")
self.assertIn("Earnings Average", data.keys(), "data missing expected key")
self.assertIn("Earnings Low", data.keys(), "data missing expected key")
self.assertIn("Earnings High", data.keys(), "data missing expected key")
self.assertIn("Revenue Average", data.keys(), "data missing expected key")
self.assertIn("Revenue Low", data.keys(), "data missing expected key")
self.assertIn("Revenue High", data.keys(), "data missing expected key")
# dividend date is not available for tested ticker GOOGL
if self.ticker.ticker != "GOOGL":
self.assertIn("Dividend Date", data.keys(), "data missing expected key")
# ex-dividend date is not always available
data_cached = self.ticker.calendar
self.assertIs(data, data_cached, "data not cached")
class TestTickerInfo(unittest.TestCase):
# Below will fail because not ported to Yahoo API
# def test_sustainability(self):
# data = self.ticker.sustainability
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.sustainability
# self.assertIs(data, data_cached, "data not cached")
# def test_shares(self):
# data = self.ticker.shares
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
class TestTickerAnalysts(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = requests_cache.CachedSession(backend='memory')
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
@@ -678,98 +712,205 @@ class TestTickerInfo(unittest.TestCase):
cls.session.close()
def setUp(self):
tkrs = ["ESLT.TA", "BP.L", "GOOGL"]
self.tickers = [yf.Ticker(tkr, session=self.session) for tkr in tkrs]
self.ticker = yf.Ticker("GOOGL", session=self.session)
def tearDown(self):
self.ticker = None
def test_recommendations(self):
data = self.ticker.recommendations
data_summary = self.ticker.recommendations_summary
self.assertTrue(data.equals(data_summary))
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.recommendations
self.assertIs(data, data_cached, "data not cached")
def test_recommendations_summary(self): # currently alias for recommendations
data = self.ticker.recommendations_summary
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
data_cached = self.ticker.recommendations_summary
self.assertIs(data, data_cached, "data not cached")
def test_upgrades_downgrades(self):
data = self.ticker.upgrades_downgrades
self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
self.assertFalse(data.empty, "data is empty")
self.assertTrue(len(data.columns) == 4, "data has wrong number of columns")
self.assertEqual(data.columns.values.tolist(), ['Firm', 'ToGrade', 'FromGrade', 'Action'], "data has wrong column names")
self.assertIsInstance(data.index, pd.DatetimeIndex, "data has wrong index type")
data_cached = self.ticker.upgrades_downgrades
self.assertIs(data, data_cached, "data not cached")
# Below will fail because not ported to Yahoo API
# def test_analyst_price_target(self):
# data = self.ticker.analyst_price_target
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.analyst_price_target
# self.assertIs(data, data_cached, "data not cached")
# def test_revenue_forecasts(self):
# data = self.ticker.revenue_forecasts
# self.assertIsInstance(data, pd.DataFrame, "data has wrong type")
# self.assertFalse(data.empty, "data is empty")
# data_cached = self.ticker.revenue_forecasts
# self.assertIs(data, data_cached, "data not cached")
class TestTickerInfo(unittest.TestCase):
session = None
@classmethod
def setUpClass(cls):
cls.session = session_gbl
@classmethod
def tearDownClass(cls):
if cls.session is not None:
cls.session.close()
def setUp(self):
self.symbols = []
self.symbols += ["ESLT.TA", "BP.L", "GOOGL"]
self.symbols.append("QCSTIX") # good for testing, doesn't trade
self.symbols += ["BTC-USD", "IWO", "VFINX", "^GSPC"]
self.symbols += ["SOKE.IS", "ADS.DE"] # detected bugs
self.tickers = [yf.Ticker(s, session=self.session) for s in self.symbols]
def tearDown(self):
self.ticker = None
def test_fast_info(self):
f = yf.Ticker("AAPL", session=self.session).fast_info
for k in f:
self.assertIsNotNone(f[k])
def test_info(self):
data = self.ticker.info
data = self.tickers[0].info
self.assertIsInstance(data, dict, "data has wrong type")
self.assertIn("symbol", data.keys(), "Did not find expected key in info dict")
self.assertEqual("GOOGL", data["symbol"], "Wrong symbol value in info dict")
expected_keys = ['industry', 'currentPrice', 'exchange', 'floatShares', 'companyOfficers', 'bid']
for k in expected_keys:
print(k)
self.assertIn("symbol", data.keys(), f"Did not find expected key '{k}' in info dict")
self.assertEqual(self.symbols[0], data["symbol"], "Wrong symbol value in info dict")
def test_basic_info(self):
yf.scrapers.quote.PRUNE_INFO = False
def test_complementary_info(self):
# This test is to check that we can successfully retrieve the trailing PEG ratio
# basic_info_keys = self.ticker.basic_info.keys()
basic_info_keys = set()
for ticker in self.tickers:
basic_info_keys.update(set(ticker.basic_info.keys()))
basic_info_keys = sorted(list(basic_info_keys))
# We don't expect this one to have a trailing PEG ratio
data1 = self.tickers[0].info
self.assertIsNone(data1['trailingPegRatio'])
key_rename_map = {}
key_rename_map["last_price"] = ["currentPrice", "regularMarketPrice"]
key_rename_map["open"] = ["open", "regularMarketOpen"]
key_rename_map["day_high"] = ["dayHigh", "regularMarketDayHigh"]
key_rename_map["day_low"] = ["dayLow", "regularMarketDayLow"]
key_rename_map["previous_close"] = ["previousClose", "regularMarketPreviousClose"]
# This one should have a trailing PEG ratio
data2 = self.tickers[2].info
self.assertIsInstance(data2['trailingPegRatio'], float)
pass
# preMarketPrice
# def test_fast_info_matches_info(self):
# fast_info_keys = set()
# for ticker in self.tickers:
# fast_info_keys.update(set(ticker.fast_info.keys()))
# fast_info_keys = sorted(list(fast_info_keys))
key_rename_map["fifty_day_average"] = "fiftyDayAverage"
key_rename_map["two_hundred_day_average"] = "twoHundredDayAverage"
key_rename_map["year_change"] = "52WeekChange"
key_rename_map["year_high"] = "fiftyTwoWeekHigh"
key_rename_map["year_low"] = "fiftyTwoWeekLow"
# key_rename_map = {}
# key_rename_map["currency"] = "currency"
# key_rename_map["quote_type"] = "quoteType"
# key_rename_map["timezone"] = "exchangeTimezoneName"
key_rename_map["last_volume"] = ["volume", "regularMarketVolume"]
key_rename_map["ten_day_average_volume"] = ["averageVolume10days", "averageDailyVolume10Day"]
key_rename_map["three_month_average_volume"] = "averageVolume"
# key_rename_map["last_price"] = ["currentPrice", "regularMarketPrice"]
# key_rename_map["open"] = ["open", "regularMarketOpen"]
# key_rename_map["day_high"] = ["dayHigh", "regularMarketDayHigh"]
# key_rename_map["day_low"] = ["dayLow", "regularMarketDayLow"]
# key_rename_map["previous_close"] = ["previousClose"]
# key_rename_map["regular_market_previous_close"] = ["regularMarketPreviousClose"]
key_rename_map["market_cap"] = "marketCap"
key_rename_map["shares"] = "floatShares"
key_rename_map["timezone"] = "exchangeTimezoneName"
# key_rename_map["fifty_day_average"] = "fiftyDayAverage"
# key_rename_map["two_hundred_day_average"] = "twoHundredDayAverage"
# key_rename_map["year_change"] = ["52WeekChange", "fiftyTwoWeekChange"]
# key_rename_map["year_high"] = "fiftyTwoWeekHigh"
# key_rename_map["year_low"] = "fiftyTwoWeekLow"
approximate_keys = {"fifty_day_average", "ten_day_average_volume"}
approximate_keys.update({"market_cap"})
# key_rename_map["last_volume"] = ["volume", "regularMarketVolume"]
# key_rename_map["ten_day_average_volume"] = ["averageVolume10days", "averageDailyVolume10Day"]
# key_rename_map["three_month_average_volume"] = "averageVolume"
# bad_keys = []
bad_keys = {"shares"}
# key_rename_map["market_cap"] = "marketCap"
# key_rename_map["shares"] = "sharesOutstanding"
# Loose tolerance for averages, no idea why don't match info[]. Is info wrong?
custom_tolerances = {}
# custom_tolerances["ten_day_average_volume"] = 1e-3
custom_tolerances["ten_day_average_volume"] = 1e-1
# custom_tolerances["three_month_average_volume"] = 1e-2
custom_tolerances["three_month_average_volume"] = 5e-1
custom_tolerances["fifty_day_average"] = 1e-2
custom_tolerances["two_hundred_day_average"] = 1e-2
# for k in list(key_rename_map.keys()):
# if '_' in k:
# key_rename_map[yf.utils.snake_case_2_camelCase(k)] = key_rename_map[k]
for k in basic_info_keys:
if k in key_rename_map:
k2 = key_rename_map[k]
else:
k2 = k
# # Note: share count items in info[] are bad. Sometimes the float > outstanding!
# # So often fast_info["shares"] does not match.
# # Why isn't fast_info["shares"] wrong? Because using it to calculate market cap always correct.
# bad_keys = {"shares"}
if not isinstance(k2, list):
k2 = [k2]
# # Loose tolerance for averages, no idea why don't match info[]. Is info wrong?
# custom_tolerances = {}
# custom_tolerances["year_change"] = 1.0
# # custom_tolerances["ten_day_average_volume"] = 1e-3
# custom_tolerances["ten_day_average_volume"] = 1e-1
# # custom_tolerances["three_month_average_volume"] = 1e-2
# custom_tolerances["three_month_average_volume"] = 5e-1
# custom_tolerances["fifty_day_average"] = 1e-2
# custom_tolerances["two_hundred_day_average"] = 1e-2
# for k in list(custom_tolerances.keys()):
# if '_' in k:
# custom_tolerances[yf.utils.snake_case_2_camelCase(k)] = custom_tolerances[k]
for m in k2:
for ticker in self.tickers:
if not m in ticker.info:
print(sorted(list(ticker.info.keys())))
raise Exception("Need to add/fix mapping for basic_info key", k)
# for k in fast_info_keys:
# if k in key_rename_map:
# k2 = key_rename_map[k]
# else:
# k2 = k
if k in bad_keys:
# Doesn't match, investigate why
continue
# if not isinstance(k2, list):
# k2 = [k2]
if k in custom_tolerances:
rtol = custom_tolerances[k]
else:
rtol = 5e-3
# rtol = 1e-4
# for m in k2:
# for ticker in self.tickers:
# if not m in ticker.info:
# # print(f"symbol={ticker.ticker}: fast_info key '{k}' mapped to info key '{m}' but not present in info")
# continue
print(f"Testing key {m} -> {k} ticker={ticker.ticker}")
# if k in approximate_keys:
v1 = ticker.basic_info[k]
v2 = ticker.info[m]
if isinstance(v1, float) or isinstance(v2, int):
self.assertTrue(np.isclose(v1, v2, rtol=rtol), f"{k}: {v1} != {v2}")
else:
self.assertEqual(v1, v2, f"{k}: {v1} != {v2}")
# if k in bad_keys:
# continue
# if k in custom_tolerances:
# rtol = custom_tolerances[k]
# else:
# rtol = 5e-3
# # rtol = 1e-4
# correct = ticker.info[m]
# test = ticker.fast_info[k]
# # print(f"Testing: symbol={ticker.ticker} m={m} k={k}: test={test} vs correct={correct}")
# if k in ["market_cap","marketCap"] and ticker.fast_info["currency"] in ["GBp", "ILA"]:
# # Adjust for currency to match Yahoo:
# test *= 0.01
# try:
# if correct is None:
# self.assertTrue(test is None or (not np.isnan(test)), f"{k}: {test} must be None or real value because correct={correct}")
# elif isinstance(test, float) or isinstance(correct, int):
# self.assertTrue(np.isclose(test, correct, rtol=rtol), f"{ticker.ticker} {k}: {test} != {correct}")
# else:
# self.assertEqual(test, correct, f"{k}: {test} != {correct}")
# except:
# if k in ["regularMarketPreviousClose"] and ticker.ticker in ["ADS.DE"]:
# # Yahoo is wrong, is returning post-market close not regular
# continue
# else:
# raise
@@ -780,6 +921,7 @@ def suite():
suite.addTest(TestTickerHolders('Test holders'))
suite.addTest(TestTickerHistory('Test Ticker history'))
suite.addTest(TestTickerMiscFinancials('Test misc financials'))
suite.addTest(TestTickerInfo('Test info & fast_info'))
return suite

92
tests/utils.py Normal file
View File

@@ -0,0 +1,92 @@
"""
Tests for utils
To run all tests in suite from commandline:
python -m unittest tests.utils
Specific test class:
python -m unittest tests.utils.TestTicker
"""
from unittest import TestSuite
# import pandas as pd
# import numpy as np
from .context import yfinance as yf
import unittest
# import requests_cache
import tempfile
import os
class TestCache(unittest.TestCase):
@classmethod
def setUpClass(cls):
cls.tempCacheDir = tempfile.TemporaryDirectory()
yf.set_tz_cache_location(cls.tempCacheDir.name)
@classmethod
def tearDownClass(cls):
cls.tempCacheDir.cleanup()
def test_storeTzNoRaise(self):
# storing TZ to cache should never raise exception
tkr = 'AMZN'
tz1 = "America/New_York"
tz2 = "London/Europe"
cache = yf.cache.get_tz_cache()
cache.store(tkr, tz1)
cache.store(tkr, tz2)
def test_setTzCacheLocation(self):
self.assertEqual(yf.cache._TzDBManager.get_location(), self.tempCacheDir.name)
tkr = 'AMZN'
tz1 = "America/New_York"
cache = yf.cache.get_tz_cache()
cache.store(tkr, tz1)
self.assertTrue(os.path.exists(os.path.join(self.tempCacheDir.name, "tkr-tz.db")))
class TestCacheNoPermission(unittest.TestCase):
@classmethod
def setUpClass(cls):
yf.set_tz_cache_location("/root/yf-cache")
def test_tzCacheRootStore(self):
# Test that if cache path in read-only filesystem, no exception.
tkr = 'AMZN'
tz1 = "America/New_York"
# During attempt to store, will discover cannot write
yf.cache.get_tz_cache().store(tkr, tz1)
# Handling the store failure replaces cache with a dummy
cache = yf.cache.get_tz_cache()
self.assertTrue(cache.dummy)
cache.store(tkr, tz1)
def test_tzCacheRootLookup(self):
# Test that if cache path in read-only filesystem, no exception.
tkr = 'AMZN'
# During attempt to lookup, will discover cannot write
yf.cache.get_tz_cache().lookup(tkr)
# Handling the lookup failure replaces cache with a dummy
cache = yf.cache.get_tz_cache()
self.assertTrue(cache.dummy)
cache.lookup(tkr)
def suite():
ts: TestSuite = unittest.TestSuite()
ts.addTest(TestCache('Test cache'))
ts.addTest(TestCacheNoPermission('Test cache no permission'))
return ts
if __name__ == '__main__':
unittest.main()

View File

@@ -23,7 +23,8 @@ from . import version
from .ticker import Ticker
from .tickers import Tickers
from .multi import download
from .utils import set_tz_cache_location
from .utils import enable_debug_mode
from .cache import set_tz_cache_location
__version__ = version.version
__author__ = "Ran Aroussi"
@@ -43,4 +44,4 @@ def pdr_override():
pass
__all__ = ['download', 'Ticker', 'Tickers', 'pdr_override', 'set_tz_cache_location']
__all__ = ['download', 'Ticker', 'Tickers', 'pdr_override', 'enable_debug_mode', 'set_tz_cache_location']

File diff suppressed because it is too large Load Diff

431
yfinance/cache.py Normal file
View File

@@ -0,0 +1,431 @@
import peewee as _peewee
from threading import Lock
import os as _os
import appdirs as _ad
import atexit as _atexit
import datetime as _datetime
import pickle as _pkl
from .utils import get_yf_logger
_cache_init_lock = Lock()
# --------------
# TimeZone cache
# --------------
class _TzCacheException(Exception):
pass
class _TzCacheDummy:
"""Dummy cache to use if tz cache is disabled"""
def lookup(self, tkr):
return None
def store(self, tkr, tz):
pass
@property
def tz_db(self):
return None
class _TzCacheManager:
_tz_cache = None
@classmethod
def get_tz_cache(cls):
if cls._tz_cache is None:
with _cache_init_lock:
cls._initialise()
return cls._tz_cache
@classmethod
def _initialise(cls, cache_dir=None):
cls._tz_cache = _TzCache()
class _TzDBManager:
_db = None
_cache_dir = _os.path.join(_ad.user_cache_dir(), "py-yfinance")
@classmethod
def get_database(cls):
if cls._db is None:
cls._initialise()
return cls._db
@classmethod
def close_db(cls):
if cls._db is not None:
try:
cls._db.close()
except Exception:
# Must discard exceptions because Python trying to quit.
pass
@classmethod
def _initialise(cls, cache_dir=None):
if cache_dir is not None:
cls._cache_dir = cache_dir
if not _os.path.isdir(cls._cache_dir):
try:
_os.makedirs(cls._cache_dir)
except OSError as err:
raise _TzCacheException(f"Error creating TzCache folder: '{cls._cache_dir}' reason: {err}")
elif not (_os.access(cls._cache_dir, _os.R_OK) and _os.access(cls._cache_dir, _os.W_OK)):
raise _TzCacheException(f"Cannot read and write in TzCache folder: '{cls._cache_dir}'")
cls._db = _peewee.SqliteDatabase(
_os.path.join(cls._cache_dir, 'tkr-tz.db'),
pragmas={'journal_mode': 'wal', 'cache_size': -64}
)
old_cache_file_path = _os.path.join(cls._cache_dir, "tkr-tz.csv")
if _os.path.isfile(old_cache_file_path):
_os.remove(old_cache_file_path)
@classmethod
def set_location(cls, new_cache_dir):
if cls._db is not None:
cls._db.close()
cls._db = None
cls._cache_dir = new_cache_dir
@classmethod
def get_location(cls):
return cls._cache_dir
# close DB when Python exists
_atexit.register(_TzDBManager.close_db)
tz_db_proxy = _peewee.Proxy()
class _KV(_peewee.Model):
key = _peewee.CharField(primary_key=True)
value = _peewee.CharField(null=True)
class Meta:
database = tz_db_proxy
without_rowid = True
class _TzCache:
def __init__(self):
self.initialised = -1
self.db = None
self.dummy = False
def get_db(self):
if self.db is not None:
return self.db
try:
self.db = _TzDBManager.get_database()
except _TzCacheException as err:
get_yf_logger().info(f"Failed to create TzCache, reason: {err}. "
"TzCache will not be used. "
"Tip: You can direct cache to use a different location with 'set_tz_cache_location(mylocation)'")
self.dummy = True
return None
return self.db
def initialise(self):
if self.initialised != -1:
return
db = self.get_db()
if db is None:
self.initialised = 0 # failure
return
db.connect()
tz_db_proxy.initialize(db)
try:
db.create_tables([_KV])
except _peewee.OperationalError as e:
if 'WITHOUT' in str(e):
_KV._meta.without_rowid = False
db.create_tables([_KV])
else:
raise
self.initialised = 1 # success
def lookup(self, key):
if self.dummy:
return None
if self.initialised == -1:
self.initialise()
if self.initialised == 0: # failure
return None
try:
return _KV.get(_KV.key == key).value
except _KV.DoesNotExist:
return None
def store(self, key, value):
if self.dummy:
return
if self.initialised == -1:
self.initialise()
if self.initialised == 0: # failure
return
db = self.get_db()
if db is None:
return
try:
if value is None:
q = _KV.delete().where(_KV.key == key)
q.execute()
return
with db.atomic():
_KV.insert(key=key, value=value).execute()
except _peewee.IntegrityError:
# Integrity error means the key already exists. Try updating the key.
old_value = self.lookup(key)
if old_value != value:
get_yf_logger().debug(f"Value for key {key} changed from {old_value} to {value}.")
with db.atomic():
q = _KV.update(value=value).where(_KV.key == key)
q.execute()
def get_tz_cache():
return _TzCacheManager.get_tz_cache()
# --------------
# Cookie cache
# --------------
class _CookieCacheException(Exception):
pass
class _CookieCacheDummy:
"""Dummy cache to use if Cookie cache is disabled"""
def lookup(self, tkr):
return None
def store(self, tkr, Cookie):
pass
@property
def Cookie_db(self):
return None
class _CookieCacheManager:
_Cookie_cache = None
@classmethod
def get_cookie_cache(cls):
if cls._Cookie_cache is None:
with _cache_init_lock:
cls._initialise()
return cls._Cookie_cache
@classmethod
def _initialise(cls, cache_dir=None):
cls._Cookie_cache = _CookieCache()
class _CookieDBManager:
_db = None
_cache_dir = _os.path.join(_ad.user_cache_dir(), "py-yfinance")
@classmethod
def get_database(cls):
if cls._db is None:
cls._initialise()
return cls._db
@classmethod
def close_db(cls):
if cls._db is not None:
try:
cls._db.close()
except Exception:
# Must discard exceptions because Python trying to quit.
pass
@classmethod
def _initialise(cls, cache_dir=None):
if cache_dir is not None:
cls._cache_dir = cache_dir
if not _os.path.isdir(cls._cache_dir):
try:
_os.makedirs(cls._cache_dir)
except OSError as err:
raise _CookieCacheException(f"Error creating CookieCache folder: '{cls._cache_dir}' reason: {err}")
elif not (_os.access(cls._cache_dir, _os.R_OK) and _os.access(cls._cache_dir, _os.W_OK)):
raise _CookieCacheException(f"Cannot read and write in CookieCache folder: '{cls._cache_dir}'")
cls._db = _peewee.SqliteDatabase(
_os.path.join(cls._cache_dir, 'cookies.db'),
pragmas={'journal_mode': 'wal', 'cache_size': -64}
)
@classmethod
def set_location(cls, new_cache_dir):
if cls._db is not None:
cls._db.close()
cls._db = None
cls._cache_dir = new_cache_dir
@classmethod
def get_location(cls):
return cls._cache_dir
# close DB when Python exists
_atexit.register(_CookieDBManager.close_db)
Cookie_db_proxy = _peewee.Proxy()
class ISODateTimeField(_peewee.DateTimeField):
# Ensure Python datetime is read & written correctly for sqlite,
# because user discovered peewee allowed an invalid datetime
# to get written.
def db_value(self, value):
if value and isinstance(value, _datetime.datetime):
return value.isoformat()
return super().db_value(value)
def python_value(self, value):
if value and isinstance(value, str) and 'T' in value:
return _datetime.datetime.fromisoformat(value)
return super().python_value(value)
class _CookieSchema(_peewee.Model):
strategy = _peewee.CharField(primary_key=True)
fetch_date = ISODateTimeField(default=_datetime.datetime.now)
# Which cookie type depends on strategy
cookie_bytes = _peewee.BlobField()
class Meta:
database = Cookie_db_proxy
without_rowid = True
class _CookieCache:
def __init__(self):
self.initialised = -1
self.db = None
self.dummy = False
def get_db(self):
if self.db is not None:
return self.db
try:
self.db = _CookieDBManager.get_database()
except _CookieCacheException as err:
get_yf_logger().info(f"Failed to create CookieCache, reason: {err}. "
"CookieCache will not be used. "
"Tip: You can direct cache to use a different location with 'set_tz_cache_location(mylocation)'")
self.dummy = True
return None
return self.db
def initialise(self):
if self.initialised != -1:
return
db = self.get_db()
if db is None:
self.initialised = 0 # failure
return
db.connect()
Cookie_db_proxy.initialize(db)
try:
db.create_tables([_CookieSchema])
except _peewee.OperationalError as e:
if 'WITHOUT' in str(e):
_CookieSchema._meta.without_rowid = False
db.create_tables([_CookieSchema])
else:
raise
self.initialised = 1 # success
def lookup(self, strategy):
if self.dummy:
return None
if self.initialised == -1:
self.initialise()
if self.initialised == 0: # failure
return None
try:
data = _CookieSchema.get(_CookieSchema.strategy == strategy)
cookie = _pkl.loads(data.cookie_bytes)
return {'cookie':cookie, 'age':_datetime.datetime.now()-data.fetch_date}
except _CookieSchema.DoesNotExist:
return None
def store(self, strategy, cookie):
if self.dummy:
return
if self.initialised == -1:
self.initialise()
if self.initialised == 0: # failure
return
db = self.get_db()
if db is None:
return
try:
q = _CookieSchema.delete().where(_CookieSchema.strategy == strategy)
q.execute()
if cookie is None:
return
with db.atomic():
cookie_pkl = _pkl.dumps(cookie, _pkl.HIGHEST_PROTOCOL)
_CookieSchema.insert(strategy=strategy, cookie_bytes=cookie_pkl).execute()
except _peewee.IntegrityError:
raise
# # Integrity error means the strategy already exists. Try updating the strategy.
# old_value = self.lookup(strategy)
# if old_value != cookie:
# get_yf_logger().debug(f"cookie for strategy {strategy} changed from {old_value} to {cookie}.")
# with db.atomic():
# q = _CookieSchema.update(cookie=cookie).where(_CookieSchema.strategy == strategy)
# q.execute()
def get_cookie_cache():
return _CookieCacheManager.get_cookie_cache()
def set_cache_location(cache_dir: str):
"""
Sets the path to create the "py-yfinance" cache folder in.
Useful if the default folder returned by "appdir.user_cache_dir()" is not writable.
Must be called before cache is used (that is, before fetching tickers).
:param cache_dir: Path to use for caches
:return: None
"""
_TzDBManager.set_location(cache_dir)
_CookieDBManager.set_location(cache_dir)
def set_tz_cache_location(cache_dir: str):
set_cache_location(cache_dir)

154
yfinance/const.py Normal file
View File

@@ -0,0 +1,154 @@
_BASE_URL_ = 'https://query2.finance.yahoo.com'
_ROOT_URL_ = 'https://finance.yahoo.com'
fundamentals_keys = {
'financials': ["TaxEffectOfUnusualItems", "TaxRateForCalcs", "NormalizedEBITDA", "NormalizedDilutedEPS",
"NormalizedBasicEPS", "TotalUnusualItems", "TotalUnusualItemsExcludingGoodwill",
"NetIncomeFromContinuingOperationNetMinorityInterest", "ReconciledDepreciation",
"ReconciledCostOfRevenue", "EBITDA", "EBIT", "NetInterestIncome", "InterestExpense",
"InterestIncome", "ContinuingAndDiscontinuedDilutedEPS", "ContinuingAndDiscontinuedBasicEPS",
"NormalizedIncome", "NetIncomeFromContinuingAndDiscontinuedOperation", "TotalExpenses",
"RentExpenseSupplemental", "ReportedNormalizedDilutedEPS", "ReportedNormalizedBasicEPS",
"TotalOperatingIncomeAsReported", "DividendPerShare", "DilutedAverageShares", "BasicAverageShares",
"DilutedEPS", "DilutedEPSOtherGainsLosses", "TaxLossCarryforwardDilutedEPS",
"DilutedAccountingChange", "DilutedExtraordinary", "DilutedDiscontinuousOperations",
"DilutedContinuousOperations", "BasicEPS", "BasicEPSOtherGainsLosses", "TaxLossCarryforwardBasicEPS",
"BasicAccountingChange", "BasicExtraordinary", "BasicDiscontinuousOperations",
"BasicContinuousOperations", "DilutedNIAvailtoComStockholders", "AverageDilutionEarnings",
"NetIncomeCommonStockholders", "OtherunderPreferredStockDividend", "PreferredStockDividends",
"NetIncome", "MinorityInterests", "NetIncomeIncludingNoncontrollingInterests",
"NetIncomeFromTaxLossCarryforward", "NetIncomeExtraordinary", "NetIncomeDiscontinuousOperations",
"NetIncomeContinuousOperations", "EarningsFromEquityInterestNetOfTax", "TaxProvision",
"PretaxIncome", "OtherIncomeExpense", "OtherNonOperatingIncomeExpenses", "SpecialIncomeCharges",
"GainOnSaleOfPPE", "GainOnSaleOfBusiness", "OtherSpecialCharges", "WriteOff",
"ImpairmentOfCapitalAssets", "RestructuringAndMergernAcquisition", "SecuritiesAmortization",
"EarningsFromEquityInterest", "GainOnSaleOfSecurity", "NetNonOperatingInterestIncomeExpense",
"TotalOtherFinanceCost", "InterestExpenseNonOperating", "InterestIncomeNonOperating",
"OperatingIncome", "OperatingExpense", "OtherOperatingExpenses", "OtherTaxes",
"ProvisionForDoubtfulAccounts", "DepreciationAmortizationDepletionIncomeStatement",
"DepletionIncomeStatement", "DepreciationAndAmortizationInIncomeStatement", "Amortization",
"AmortizationOfIntangiblesIncomeStatement", "DepreciationIncomeStatement", "ResearchAndDevelopment",
"SellingGeneralAndAdministration", "SellingAndMarketingExpense", "GeneralAndAdministrativeExpense",
"OtherGandA", "InsuranceAndClaims", "RentAndLandingFees", "SalariesAndWages", "GrossProfit",
"CostOfRevenue", "TotalRevenue", "ExciseTaxes", "OperatingRevenue"],
'balance-sheet': ["TreasurySharesNumber", "PreferredSharesNumber", "OrdinarySharesNumber", "ShareIssued", "NetDebt",
"TotalDebt", "TangibleBookValue", "InvestedCapital", "WorkingCapital", "NetTangibleAssets",
"CapitalLeaseObligations", "CommonStockEquity", "PreferredStockEquity", "TotalCapitalization",
"TotalEquityGrossMinorityInterest", "MinorityInterest", "StockholdersEquity",
"OtherEquityInterest", "GainsLossesNotAffectingRetainedEarnings", "OtherEquityAdjustments",
"FixedAssetsRevaluationReserve", "ForeignCurrencyTranslationAdjustments",
"MinimumPensionLiabilities", "UnrealizedGainLoss", "TreasuryStock", "RetainedEarnings",
"AdditionalPaidInCapital", "CapitalStock", "OtherCapitalStock", "CommonStock", "PreferredStock",
"TotalPartnershipCapital", "GeneralPartnershipCapital", "LimitedPartnershipCapital",
"TotalLiabilitiesNetMinorityInterest", "TotalNonCurrentLiabilitiesNetMinorityInterest",
"OtherNonCurrentLiabilities", "LiabilitiesHeldforSaleNonCurrent", "RestrictedCommonStock",
"PreferredSecuritiesOutsideStockEquity", "DerivativeProductLiabilities", "EmployeeBenefits",
"NonCurrentPensionAndOtherPostretirementBenefitPlans", "NonCurrentAccruedExpenses",
"DuetoRelatedPartiesNonCurrent", "TradeandOtherPayablesNonCurrent",
"NonCurrentDeferredLiabilities", "NonCurrentDeferredRevenue",
"NonCurrentDeferredTaxesLiabilities", "LongTermDebtAndCapitalLeaseObligation",
"LongTermCapitalLeaseObligation", "LongTermDebt", "LongTermProvisions", "CurrentLiabilities",
"OtherCurrentLiabilities", "CurrentDeferredLiabilities", "CurrentDeferredRevenue",
"CurrentDeferredTaxesLiabilities", "CurrentDebtAndCapitalLeaseObligation",
"CurrentCapitalLeaseObligation", "CurrentDebt", "OtherCurrentBorrowings", "LineOfCredit",
"CommercialPaper", "CurrentNotesPayable", "PensionandOtherPostRetirementBenefitPlansCurrent",
"CurrentProvisions", "PayablesAndAccruedExpenses", "CurrentAccruedExpenses", "InterestPayable",
"Payables", "OtherPayable", "DuetoRelatedPartiesCurrent", "DividendsPayable", "TotalTaxPayable",
"IncomeTaxPayable", "AccountsPayable", "TotalAssets", "TotalNonCurrentAssets",
"OtherNonCurrentAssets", "DefinedPensionBenefit", "NonCurrentPrepaidAssets",
"NonCurrentDeferredAssets", "NonCurrentDeferredTaxesAssets", "DuefromRelatedPartiesNonCurrent",
"NonCurrentNoteReceivables", "NonCurrentAccountsReceivable", "FinancialAssets",
"InvestmentsAndAdvances", "OtherInvestments", "InvestmentinFinancialAssets",
"HeldToMaturitySecurities", "AvailableForSaleSecurities",
"FinancialAssetsDesignatedasFairValueThroughProfitorLossTotal", "TradingSecurities",
"LongTermEquityInvestment", "InvestmentsinJointVenturesatCost",
"InvestmentsInOtherVenturesUnderEquityMethod", "InvestmentsinAssociatesatCost",
"InvestmentsinSubsidiariesatCost", "InvestmentProperties", "GoodwillAndOtherIntangibleAssets",
"OtherIntangibleAssets", "Goodwill", "NetPPE", "AccumulatedDepreciation", "GrossPPE", "Leases",
"ConstructionInProgress", "OtherProperties", "MachineryFurnitureEquipment",
"BuildingsAndImprovements", "LandAndImprovements", "Properties", "CurrentAssets",
"OtherCurrentAssets", "HedgingAssetsCurrent", "AssetsHeldForSaleCurrent", "CurrentDeferredAssets",
"CurrentDeferredTaxesAssets", "RestrictedCash", "PrepaidAssets", "Inventory",
"InventoriesAdjustmentsAllowances", "OtherInventories", "FinishedGoods", "WorkInProcess",
"RawMaterials", "Receivables", "ReceivablesAdjustmentsAllowances", "OtherReceivables",
"DuefromRelatedPartiesCurrent", "TaxesReceivable", "AccruedInterestReceivable", "NotesReceivable",
"LoansReceivable", "AccountsReceivable", "AllowanceForDoubtfulAccountsReceivable",
"GrossAccountsReceivable", "CashCashEquivalentsAndShortTermInvestments",
"OtherShortTermInvestments", "CashAndCashEquivalents", "CashEquivalents", "CashFinancial"],
'cash-flow': ["ForeignSales", "DomesticSales", "AdjustedGeographySegmentData", "FreeCashFlow",
"RepurchaseOfCapitalStock", "RepaymentOfDebt", "IssuanceOfDebt", "IssuanceOfCapitalStock",
"CapitalExpenditure", "InterestPaidSupplementalData", "IncomeTaxPaidSupplementalData",
"EndCashPosition", "OtherCashAdjustmentOutsideChangeinCash", "BeginningCashPosition",
"EffectOfExchangeRateChanges", "ChangesInCash", "OtherCashAdjustmentInsideChangeinCash",
"CashFlowFromDiscontinuedOperation", "FinancingCashFlow", "CashFromDiscontinuedFinancingActivities",
"CashFlowFromContinuingFinancingActivities", "NetOtherFinancingCharges", "InterestPaidCFF",
"ProceedsFromStockOptionExercised", "CashDividendsPaid", "PreferredStockDividendPaid",
"CommonStockDividendPaid", "NetPreferredStockIssuance", "PreferredStockPayments",
"PreferredStockIssuance", "NetCommonStockIssuance", "CommonStockPayments", "CommonStockIssuance",
"NetIssuancePaymentsOfDebt", "NetShortTermDebtIssuance", "ShortTermDebtPayments",
"ShortTermDebtIssuance", "NetLongTermDebtIssuance", "LongTermDebtPayments", "LongTermDebtIssuance",
"InvestingCashFlow", "CashFromDiscontinuedInvestingActivities",
"CashFlowFromContinuingInvestingActivities", "NetOtherInvestingChanges", "InterestReceivedCFI",
"DividendsReceivedCFI", "NetInvestmentPurchaseAndSale", "SaleOfInvestment", "PurchaseOfInvestment",
"NetInvestmentPropertiesPurchaseAndSale", "SaleOfInvestmentProperties",
"PurchaseOfInvestmentProperties", "NetBusinessPurchaseAndSale", "SaleOfBusiness",
"PurchaseOfBusiness", "NetIntangiblesPurchaseAndSale", "SaleOfIntangibles", "PurchaseOfIntangibles",
"NetPPEPurchaseAndSale", "SaleOfPPE", "PurchaseOfPPE", "CapitalExpenditureReported",
"OperatingCashFlow", "CashFromDiscontinuedOperatingActivities",
"CashFlowFromContinuingOperatingActivities", "TaxesRefundPaid", "InterestReceivedCFO",
"InterestPaidCFO", "DividendReceivedCFO", "DividendPaidCFO", "ChangeInWorkingCapital",
"ChangeInOtherWorkingCapital", "ChangeInOtherCurrentLiabilities", "ChangeInOtherCurrentAssets",
"ChangeInPayablesAndAccruedExpense", "ChangeInAccruedExpense", "ChangeInInterestPayable",
"ChangeInPayable", "ChangeInDividendPayable", "ChangeInAccountPayable", "ChangeInTaxPayable",
"ChangeInIncomeTaxPayable", "ChangeInPrepaidAssets", "ChangeInInventory", "ChangeInReceivables",
"ChangesInAccountReceivables", "OtherNonCashItems", "ExcessTaxBenefitFromStockBasedCompensation",
"StockBasedCompensation", "UnrealizedGainLossOnInvestmentSecurities", "ProvisionandWriteOffofAssets",
"AssetImpairmentCharge", "AmortizationOfSecurities", "DeferredTax", "DeferredIncomeTax",
"DepreciationAmortizationDepletion", "Depletion", "DepreciationAndAmortization",
"AmortizationCashFlow", "AmortizationOfIntangibles", "Depreciation", "OperatingGainsLosses",
"PensionAndEmployeeBenefitExpense", "EarningsLossesFromEquityInvestments",
"GainLossOnInvestmentSecurities", "NetForeignCurrencyExchangeGainLoss", "GainLossOnSaleOfPPE",
"GainLossOnSaleOfBusiness", "NetIncomeFromContinuingOperations",
"CashFlowsfromusedinOperatingActivitiesDirect", "TaxesRefundPaidDirect", "InterestReceivedDirect",
"InterestPaidDirect", "DividendsReceivedDirect", "DividendsPaidDirect", "ClassesofCashPayments",
"OtherCashPaymentsfromOperatingActivities", "PaymentsonBehalfofEmployees",
"PaymentstoSuppliersforGoodsandServices", "ClassesofCashReceiptsfromOperatingActivities",
"OtherCashReceiptsfromOperatingActivities", "ReceiptsfromGovernmentGrants", "ReceiptsfromCustomers"]}
_PRICE_COLNAMES_ = ['Open', 'High', 'Low', 'Close', 'Adj Close']
quote_summary_valid_modules = (
"summaryProfile", # contains general information about the company
"summaryDetail", # prices + volume + market cap + etc
"assetProfile", # summaryProfile + company officers
"fundProfile",
"price", # current prices
"quoteType", # quoteType
"esgScores", # Environmental, social, and governance (ESG) scores, sustainability and ethical performance of companies
"incomeStatementHistory",
"incomeStatementHistoryQuarterly",
"balanceSheetHistory",
"balanceSheetHistoryQuarterly",
"cashFlowStatementHistory",
"cashFlowStatementHistoryQuarterly",
"defaultKeyStatistics", # KPIs (PE, enterprise value, EPS, EBITA, and more)
"financialData", # Financial KPIs (revenue, gross margins, operating cash flow, free cash flow, and more)
"calendarEvents", # future earnings date
"secFilings", # SEC filings, such as 10K and 10Q reports
"upgradeDowngradeHistory", # upgrades and downgrades that analysts have given a company's stock
"institutionOwnership", # institutional ownership, holders and shares outstanding
"fundOwnership", # mutual fund ownership, holders and shares outstanding
"majorDirectHolders",
"majorHoldersBreakdown",
"insiderTransactions", # insider transactions, such as the number of shares bought and sold by company executives
"insiderHolders", # insider holders, such as the number of shares held by company executives
"netSharePurchaseActivity", # net share purchase activity, such as the number of shares bought and sold by company executives
"earnings", # earnings history
"earningsHistory",
"earningsTrend", # earnings trend
"industryTrend",
"indexTrend",
"sectorTrend",
"recommendationTrend",
"futuresChain",
)

View File

@@ -1,26 +1,14 @@
import functools
from functools import lru_cache
import hashlib
from base64 import b64decode
usePycryptodome = False # slightly faster
# usePycryptodome = True
if usePycryptodome:
from Crypto.Cipher import AES
from Crypto.Util.Padding import unpad
else:
from cryptography.hazmat.primitives import padding
from cryptography.hazmat.primitives.ciphers import Cipher, algorithms, modes
import requests as requests
import re
from bs4 import BeautifulSoup
import datetime
from frozendict import frozendict
try:
import ujson as json
except ImportError:
import json as json
from . import utils, cache
import threading
cache_maxsize = 64
@@ -46,175 +34,351 @@ def lru_cache_freezeargs(func):
return wrapped
def decrypt_cryptojs_aes_stores(data):
encrypted_stores = data['context']['dispatcher']['stores']
password = None
candidate_passwords = []
if "_cs" in data and "_cr" in data:
_cs = data["_cs"]
_cr = data["_cr"]
_cr = b"".join(int.to_bytes(i, length=4, byteorder="big", signed=True) for i in json.loads(_cr)["words"])
password = hashlib.pbkdf2_hmac("sha1", _cs.encode("utf8"), _cr, 1, dklen=32).hex()
else:
# Currently assume one extra key in dict, which is password. Print error if
# more extra keys detected.
new_keys = [k for k in data.keys() if k not in ["context", "plugins"]]
new_keys_values = set([data[k] for k in new_keys])
# Maybe multiple keys have same value - keep one of each
new_keys2 = []
new_keys2_values = set()
for k in new_keys:
v = data[k]
if not v in new_keys2_values:
new_keys2.append(k)
new_keys2_values.add(v)
l = len(new_keys)
if l == 0:
return None
elif l == 1 and isinstance(data[new_keys[0]], str):
password_key = new_keys[0]
# else:
# msg = "Yahoo has again changed data format, yfinance now unsure which key(s) is for decryption:"
# new_keys_pretty = {}
# l = min(10, len(new_keys))
# for i in range(0, l):
# k = new_keys[i]
# k_str = k if len(k) < 32 else k[:32-3]+"..."
# v = data[k]
# v_type = type(v)
# v_str = str(v)
# if len(v_str) > 256:
# v_str = v_str[:256]+"..."
# new_keys_pretty[k_str] = f"{v_str}' ({v_type})"
# for k in new_keys_pretty:
# msg += '\n' + f"'{k}' -> '{new_keys_pretty[k]}'"
# if len(new_keys) > l:
# d = len(new_keys) - l
# msg += '\n' + "..."
# msg += '\n' + f"{d} more options!"
# raise Exception(msg)
# password_key = new_keys[0]
# password = data[password_key]
# The above attempt to smartly pick out decryption key has stopped working.
# Fortunately the keys Yahoo use are currently hardcoded in their JSON:
candidate_passwords += ["ad4d90b3c9f2e1d156ef98eadfa0ff93e4042f6960e54aa2a13f06f528e6b50ba4265a26a1fd5b9cd3db0d268a9c34e1d080592424309429a58bce4adc893c87", \
"e9a8ab8e5620b712ebc2fb4f33d5c8b9c80c0d07e8c371911c785cf674789f1747d76a909510158a7b7419e86857f2d7abbd777813ff64840e4cbc514d12bcae",
"6ae2523aeafa283dad746556540145bf603f44edbf37ad404d3766a8420bb5eb1d3738f52a227b88283cca9cae44060d5f0bba84b6a495082589f5fe7acbdc9e",
"3365117c2a368ffa5df7313a4a84988f73926a86358e8eea9497c5ff799ce27d104b68e5f2fbffa6f8f92c1fef41765a7066fa6bcf050810a9c4c7872fd3ebf0"]
# candidate_passwords += [data[k] for k in new_keys] # don't do these, none work
encrypted_stores = b64decode(encrypted_stores)
assert encrypted_stores[0:8] == b"Salted__"
salt = encrypted_stores[8:16]
encrypted_stores = encrypted_stores[16:]
def _EVPKDF(password, salt, keySize=32, ivSize=16, iterations=1, hashAlgorithm="md5") -> tuple:
"""OpenSSL EVP Key Derivation Function
Args:
password (Union[str, bytes, bytearray]): Password to generate key from.
salt (Union[bytes, bytearray]): Salt to use.
keySize (int, optional): Output key length in bytes. Defaults to 32.
ivSize (int, optional): Output Initialization Vector (IV) length in bytes. Defaults to 16.
iterations (int, optional): Number of iterations to perform. Defaults to 1.
hashAlgorithm (str, optional): Hash algorithm to use for the KDF. Defaults to 'md5'.
Returns:
key, iv: Derived key and Initialization Vector (IV) bytes.
Taken from: https://gist.github.com/rafiibrahim8/0cd0f8c46896cafef6486cb1a50a16d3
OpenSSL original code: https://github.com/openssl/openssl/blob/master/crypto/evp/evp_key.c#L78
"""
assert iterations > 0, "Iterations can not be less than 1."
if isinstance(password, str):
password = password.encode("utf-8")
final_length = keySize + ivSize
key_iv = b""
block = None
while len(key_iv) < final_length:
hasher = hashlib.new(hashAlgorithm)
if block:
hasher.update(block)
hasher.update(password)
hasher.update(salt)
block = hasher.digest()
for _ in range(1, iterations):
block = hashlib.new(hashAlgorithm, block).digest()
key_iv += block
key, iv = key_iv[:keySize], key_iv[keySize:final_length]
return key, iv
def _decrypt(encrypted_stores, password, key, iv):
if usePycryptodome:
cipher = AES.new(key, AES.MODE_CBC, iv=iv)
plaintext = cipher.decrypt(encrypted_stores)
plaintext = unpad(plaintext, 16, style="pkcs7")
else:
cipher = Cipher(algorithms.AES(key), modes.CBC(iv))
decryptor = cipher.decryptor()
plaintext = decryptor.update(encrypted_stores) + decryptor.finalize()
unpadder = padding.PKCS7(128).unpadder()
plaintext = unpadder.update(plaintext) + unpadder.finalize()
plaintext = plaintext.decode("utf-8")
return plaintext
if not password is None:
try:
key, iv = _EVPKDF(password, salt, keySize=32, ivSize=16, iterations=1, hashAlgorithm="md5")
except:
raise Exception("yfinance failed to decrypt Yahoo data response")
plaintext = _decrypt(encrypted_stores, password, key, iv)
else:
success = False
for i in range(len(candidate_passwords)):
# print(f"Trying candiate pw {i+1}/{len(candidate_passwords)}")
password = candidate_passwords[i]
try:
key, iv = _EVPKDF(password, salt, keySize=32, ivSize=16, iterations=1, hashAlgorithm="md5")
plaintext = _decrypt(encrypted_stores, password, key, iv)
success = True
break
except:
pass
if not success:
raise Exception("yfinance failed to decrypt Yahoo data response with hardcoded keys, contact developers")
decoded_stores = json.loads(plaintext)
return decoded_stores
_SCRAPE_URL_ = 'https://finance.yahoo.com/quote'
class TickerData:
class SingletonMeta(type):
"""
Have one place to retrieve data from Yahoo API in order to ease caching and speed up operations
Metaclass that creates a Singleton instance.
"""
_instances = {}
_lock = threading.Lock()
def __call__(cls, *args, **kwargs):
with cls._lock:
if cls not in cls._instances:
instance = super().__call__(*args, **kwargs)
cls._instances[cls] = instance
else:
cls._instances[cls]._set_session(*args, **kwargs)
return cls._instances[cls]
class YfData(metaclass=SingletonMeta):
"""
Have one place to retrieve data from Yahoo API in order to ease caching and speed up operations.
Singleton means one session one cookie shared by all threads.
"""
user_agent_headers = {
'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/39.0.2171.95 Safari/537.36'}
def __init__(self, ticker: str, session=None):
self.ticker = ticker
self._session = session or requests
def __init__(self, session=None):
self._session = session or requests.Session()
def get(self, url, user_agent_headers=None, params=None, proxy=None, timeout=30):
proxy = self._get_proxy(proxy)
try:
self._session.cache
except AttributeError:
# Not caching
self._session_is_caching = False
else:
# Is caching. This is annoying.
# Can't simply use a non-caching session to fetch cookie & crumb,
# because then the caching-session won't have cookie.
self._session_is_caching = True
from requests_cache import DO_NOT_CACHE
self._expire_after = DO_NOT_CACHE
self._crumb = None
self._cookie = None
if self._session_is_caching and self._cookie is None:
utils.print_once("WARNING: cookie & crumb does not work well with requests_cache. Am experimenting with 'expire_after=DO_NOT_CACHE', but you need to help stress-test.")
# Default to using 'basic' strategy
self._cookie_strategy = 'basic'
# If it fails, then fallback method is 'csrf'
# self._cookie_strategy = 'csrf'
self._cookie_lock = threading.Lock()
def _set_session(self, session):
if session is None:
return
with self._cookie_lock:
self._session = session
def _set_cookie_strategy(self, strategy, have_lock=False):
if strategy == self._cookie_strategy:
return
if not have_lock:
self._cookie_lock.acquire()
try:
if self._cookie_strategy == 'csrf':
utils.get_yf_logger().debug(f'toggling cookie strategy {self._cookie_strategy} -> basic')
self._session.cookies.clear()
self._cookie_strategy = 'basic'
else:
utils.get_yf_logger().debug(f'toggling cookie strategy {self._cookie_strategy} -> csrf')
self._cookie_strategy = 'csrf'
self._cookie = None
self._crumb = None
except Exception:
self._cookie_lock.release()
raise
if not have_lock:
self._cookie_lock.release()
def _save_session_cookies(self):
try:
cache.get_cookie_cache().store('csrf', self._session.cookies)
except Exception:
return False
return True
def _load_session_cookies(self):
cookie_dict = cache.get_cookie_cache().lookup('csrf')
if cookie_dict is None:
return False
# Periodically refresh, 24 hours seems fair.
if cookie_dict['age'] > datetime.timedelta(days=1):
return False
self._session.cookies.update(cookie_dict['cookie'])
utils.get_yf_logger().debug('loaded persistent cookie')
def _save_cookie_basic(self, cookie):
try:
cache.get_cookie_cache().store('basic', cookie)
except Exception:
return False
return True
def _load_cookie_basic(self):
cookie_dict = cache.get_cookie_cache().lookup('basic')
if cookie_dict is None:
return None
# Periodically refresh, 24 hours seems fair.
if cookie_dict['age'] > datetime.timedelta(days=1):
return None
utils.get_yf_logger().debug('loaded persistent cookie')
return cookie_dict['cookie']
def _get_cookie_basic(self, proxy=None, timeout=30):
if self._cookie is not None:
utils.get_yf_logger().debug('reusing cookie')
return self._cookie
self._cookie = self._load_cookie_basic()
if self._cookie is not None:
return self._cookie
# To avoid infinite recursion, do NOT use self.get()
# - 'allow_redirects' copied from @psychoz971 solution - does it help USA?
response = self._session.get(
url=url,
params=params,
url='https://fc.yahoo.com',
headers=self.user_agent_headers,
proxies=proxy,
timeout=timeout,
headers=user_agent_headers or self.user_agent_headers)
allow_redirects=True)
if not response.cookies:
utils.get_yf_logger().debug("response.cookies = None")
return None
self._cookie = list(response.cookies)[0]
if self._cookie == '':
utils.get_yf_logger().debug("list(response.cookies)[0] = ''")
return None
self._save_cookie_basic(self._cookie)
utils.get_yf_logger().debug(f"fetched basic cookie = {self._cookie}")
return self._cookie
def _get_crumb_basic(self, proxy=None, timeout=30):
if self._crumb is not None:
utils.get_yf_logger().debug('reusing crumb')
return self._crumb
cookie = self._get_cookie_basic()
if cookie is None:
return None
# - 'allow_redirects' copied from @psychoz971 solution - does it help USA?
get_args = {
'url': "https://query1.finance.yahoo.com/v1/test/getcrumb",
'headers': self.user_agent_headers,
'cookies': {cookie.name: cookie.value},
'proxies': proxy,
'timeout': timeout,
'allow_redirects': True
}
if self._session_is_caching:
get_args['expire_after'] = self._expire_after
crumb_response = self._session.get(**get_args)
else:
crumb_response = self._session.get(**get_args)
self._crumb = crumb_response.text
if self._crumb is None or '<html>' in self._crumb:
utils.get_yf_logger().debug("Didn't receive crumb")
return None
utils.get_yf_logger().debug(f"crumb = '{self._crumb}'")
return self._crumb
@utils.log_indent_decorator
def _get_cookie_and_crumb_basic(self, proxy, timeout):
cookie = self._get_cookie_basic(proxy, timeout)
crumb = self._get_crumb_basic(proxy, timeout)
return cookie, crumb
def _get_cookie_csrf(self, proxy, timeout):
if self._cookie is not None:
utils.get_yf_logger().debug('reusing cookie')
return True
elif self._load_session_cookies():
utils.get_yf_logger().debug('reusing persistent cookie')
self._cookie = True
return True
base_args = {
'headers': self.user_agent_headers,
'proxies': proxy,
'timeout': timeout}
get_args = {**base_args, 'url': 'https://guce.yahoo.com/consent'}
if self._session_is_caching:
get_args['expire_after'] = self._expire_after
response = self._session.get(**get_args)
else:
response = self._session.get(**get_args)
soup = BeautifulSoup(response.content, 'html.parser')
csrfTokenInput = soup.find('input', attrs={'name': 'csrfToken'})
if csrfTokenInput is None:
utils.get_yf_logger().debug('Failed to find "csrfToken" in response')
return False
csrfToken = csrfTokenInput['value']
utils.get_yf_logger().debug(f'csrfToken = {csrfToken}')
sessionIdInput = soup.find('input', attrs={'name': 'sessionId'})
sessionId = sessionIdInput['value']
utils.get_yf_logger().debug(f"sessionId='{sessionId}")
originalDoneUrl = 'https://finance.yahoo.com/'
namespace = 'yahoo'
data = {
'agree': ['agree', 'agree'],
'consentUUID': 'default',
'sessionId': sessionId,
'csrfToken': csrfToken,
'originalDoneUrl': originalDoneUrl,
'namespace': namespace,
}
post_args = {**base_args,
'url': f'https://consent.yahoo.com/v2/collectConsent?sessionId={sessionId}',
'data': data}
get_args = {**base_args,
'url': f'https://guce.yahoo.com/copyConsent?sessionId={sessionId}',
'data': data}
if self._session_is_caching:
post_args['expire_after'] = self._expire_after
get_args['expire_after'] = self._expire_after
self._session.post(**post_args)
self._session.get(**get_args)
else:
self._session.post(**post_args)
self._session.get(**get_args)
self._cookie = True
self._save_session_cookies()
return True
@utils.log_indent_decorator
def _get_crumb_csrf(self, proxy=None, timeout=30):
# Credit goes to @bot-unit #1729
if self._crumb is not None:
utils.get_yf_logger().debug('reusing crumb')
return self._crumb
if not self._get_cookie_csrf(proxy, timeout):
# This cookie stored in session
return None
get_args = {
'url': 'https://query2.finance.yahoo.com/v1/test/getcrumb',
'headers': self.user_agent_headers,
'proxies': proxy,
'timeout': timeout}
if self._session_is_caching:
get_args['expire_after'] = self._expire_after
r = self._session.get(**get_args)
else:
r = self._session.get(**get_args)
self._crumb = r.text
if self._crumb is None or '<html>' in self._crumb or self._crumb == '':
utils.get_yf_logger().debug("Didn't receive crumb")
return None
utils.get_yf_logger().debug(f"crumb = '{self._crumb}'")
return self._crumb
@utils.log_indent_decorator
def _get_cookie_and_crumb(self, proxy=None, timeout=30):
cookie, crumb, strategy = None, None, None
utils.get_yf_logger().debug(f"cookie_mode = '{self._cookie_strategy}'")
with self._cookie_lock:
if self._cookie_strategy == 'csrf':
crumb = self._get_crumb_csrf()
if crumb is None:
# Fail
self._set_cookie_strategy('basic', have_lock=True)
cookie, crumb = self._get_cookie_and_crumb_basic(proxy, timeout)
else:
# Fallback strategy
cookie, crumb = self._get_cookie_and_crumb_basic(proxy, timeout)
if cookie is None or crumb is None:
# Fail
self._set_cookie_strategy('csrf', have_lock=True)
crumb = self._get_crumb_csrf()
strategy = self._cookie_strategy
return cookie, crumb, strategy
@utils.log_indent_decorator
def get(self, url, user_agent_headers=None, params=None, proxy=None, timeout=30):
# Important: treat input arguments as immutable.
if len(url) > 200:
utils.get_yf_logger().debug(f'url={url[:200]}...')
else:
utils.get_yf_logger().debug(f'url={url}')
utils.get_yf_logger().debug(f'params={params}')
proxy = self._get_proxy(proxy)
if params is None:
params = {}
if 'crumb' in params:
raise Exception("Don't manually add 'crumb' to params dict, let data.py handle it")
cookie, crumb, strategy = self._get_cookie_and_crumb()
if crumb is not None:
crumbs = {'crumb': crumb}
else:
crumbs = {}
if strategy == 'basic' and cookie is not None:
# Basic cookie strategy adds cookie to GET parameters
cookies = {cookie.name: cookie.value}
else:
cookies = None
request_args = {
'url': url,
'params': {**params, **crumbs},
'cookies': cookies,
'proxies': proxy,
'timeout': timeout,
'headers': user_agent_headers or self.user_agent_headers
}
response = self._session.get(**request_args)
utils.get_yf_logger().debug(f'response code={response.status_code}')
if response.status_code >= 400:
# Retry with other cookie strategy
if strategy == 'basic':
self._set_cookie_strategy('csrf')
else:
self._set_cookie_strategy('basic')
cookie, crumb, strategy = self._get_cookie_and_crumb(proxy, timeout)
request_args['params']['crumb'] = crumb
if strategy == 'basic':
request_args['cookies'] = {cookie.name: cookie.value}
response = self._session.get(**request_args)
utils.get_yf_logger().debug(f'response code={response.status_code}')
return response
@lru_cache_freezeargs
@@ -225,45 +389,13 @@ class TickerData:
def _get_proxy(self, proxy):
# setup proxy in requests format
if proxy is not None:
if isinstance(proxy, dict) and "https" in proxy:
if isinstance(proxy, (dict, frozendict)) and "https" in proxy:
proxy = proxy["https"]
proxy = {"https": proxy}
return proxy
@lru_cache_freezeargs
@lru_cache(maxsize=cache_maxsize)
def get_json_data_stores(self, sub_page: str = None, proxy=None) -> dict:
'''
get_json_data_stores returns a python dictionary of the data stores in yahoo finance web page.
'''
if sub_page:
ticker_url = "{}/{}/{}".format(_SCRAPE_URL_, self.ticker, sub_page)
else:
ticker_url = "{}/{}".format(_SCRAPE_URL_, self.ticker)
html = self.get(url=ticker_url, proxy=proxy).text
# The actual json-data for stores is in a javascript assignment in the webpage
try:
json_str = html.split('root.App.main =')[1].split(
'(this)')[0].split(';\n}')[0].strip()
except IndexError:
# Fetch failed, probably because Yahoo spam triggered
return {}
data = json.loads(json_str)
stores = decrypt_cryptojs_aes_stores(data)
if stores is None:
# Maybe Yahoo returned old format, not encrypted
if "context" in data and "dispatcher" in data["context"]:
stores = data['context']['dispatcher']['stores']
if stores is None:
raise Exception(f"{self.ticker}: Failed to extract data stores from web request")
# return data
new_data = json.dumps(stores).replace('{}', 'null')
new_data = re.sub(
r'{[\'|\"]raw[\'|\"]:(.*?),(.*?)}', r'\1', new_data)
return json.loads(new_data)
def get_raw_json(self, url, user_agent_headers=None, params=None, proxy=None, timeout=30):
utils.get_yf_logger().debug(f'get_raw_json(): {url}')
response = self.get(url, user_agent_headers=user_agent_headers, params=params, proxy=proxy, timeout=timeout)
response.raise_for_status()
return response.json()

View File

@@ -4,3 +4,9 @@ class YFinanceException(Exception):
class YFinanceDataException(YFinanceException):
pass
class YFNotImplementedError(NotImplementedError):
def __init__(self, method_name):
super().__init__(f"Have not implemented fetching '{method_name}' from Yahoo API")

View File

@@ -21,18 +21,23 @@
from __future__ import print_function
import logging
import time as _time
import traceback
import multitasking as _multitasking
import pandas as _pd
from . import Ticker, utils
from .data import YfData
from . import shared
def download(tickers, start=None, end=None, actions=False, threads=True, ignore_tz=False,
@utils.log_indent_decorator
def download(tickers, start=None, end=None, actions=False, threads=True, ignore_tz=None,
group_by='column', auto_adjust=False, back_adjust=False, repair=False, keepna=False,
progress=True, period="max", show_errors=True, interval="1d", prepost=False,
proxy=None, rounding=False, timeout=10):
progress=True, period="max", show_errors=None, interval="1d", prepost=False,
proxy=None, rounding=False, timeout=10, session=None):
"""Download yahoo tickers
:Parameters:
tickers : str, list
@@ -44,11 +49,13 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
Valid intervals: 1m,2m,5m,15m,30m,60m,90m,1h,1d,5d,1wk,1mo,3mo
Intraday data cannot extend last 60 days
start: str
Download start date string (YYYY-MM-DD) or _datetime.
Default is 1900-01-01
Download start date string (YYYY-MM-DD) or _datetime, inclusive.
Default is 99 years ago
E.g. for start="2020-01-01", the first data point will be on "2020-01-01"
end: str
Download end date string (YYYY-MM-DD) or _datetime.
Download end date string (YYYY-MM-DD) or _datetime, exclusive.
Default is now
E.g. for end="2023-01-01", the last data point will be on "2022-12-31"
group_by : str
Group by 'ticker' or 'column' (default)
prepost : bool
@@ -68,17 +75,48 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
How many threads to use for mass downloading. Default is True
ignore_tz: bool
When combining from different timezones, ignore that part of datetime.
Default is False
Default depends on interval. Intraday = False. Day+ = True.
proxy: str
Optional. Proxy server URL scheme. Default is None
rounding: bool
Optional. Round values to 2 decimal places?
show_errors: bool
Optional. Doesn't print errors if False
DEPRECATED, will be removed in future version
timeout: None or float
If not None stops waiting for a response after given number of
seconds. (Can also be a fraction of a second e.g. 0.01)
session: None or Session
Optional. Pass your own session object to be used for all requests
"""
logger = utils.get_yf_logger()
if show_errors is not None:
if show_errors:
utils.print_once(f"yfinance: download(show_errors={show_errors}) argument is deprecated and will be removed in future version. Do this instead: logging.getLogger('yfinance').setLevel(logging.ERROR)")
logger.setLevel(logging.ERROR)
else:
utils.print_once(f"yfinance: download(show_errors={show_errors}) argument is deprecated and will be removed in future version. Do this instead to suppress error messages: logging.getLogger('yfinance').setLevel(logging.CRITICAL)")
logger.setLevel(logging.CRITICAL)
if logger.isEnabledFor(logging.DEBUG):
if threads:
# With DEBUG, each thread generates a lot of log messages.
# And with multi-threading, these messages will be interleaved, bad!
# So disable multi-threading to make log readable.
logger.debug('Disabling multithreading because DEBUG logging enabled')
threads = False
if progress:
# Disable progress bar, interferes with display of log messages
progress = False
if ignore_tz is None:
# Set default value depending on interval
if interval[1:] in ['m', 'h']:
# Intraday
ignore_tz = False
else:
ignore_tz = True
# create ticker list
tickers = tickers if isinstance(
@@ -90,7 +128,7 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
for ticker in tickers:
if utils.is_isin(ticker):
isin = ticker
ticker = utils.get_ticker_by_isin(ticker, proxy)
ticker = utils.get_ticker_by_isin(ticker, proxy, session=session)
shared._ISINS[ticker] = isin
_tickers_.append(ticker)
@@ -104,6 +142,10 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
# reset shared._DFS
shared._DFS = {}
shared._ERRORS = {}
shared._TRACEBACKS = {}
# Ensure data initialised with session.
YfData(session=session)
# download using threads
if threads:
@@ -119,7 +161,6 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
rounding=rounding, timeout=timeout)
while len(shared._DFS) < len(tickers):
_time.sleep(0.01)
# download synchronously
else:
for i, ticker in enumerate(tickers):
@@ -129,19 +170,41 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
back_adjust=back_adjust, repair=repair, keepna=keepna,
proxy=proxy,
rounding=rounding, timeout=timeout)
shared._DFS[ticker.upper()] = data
if progress:
shared._PROGRESS_BAR.animate()
if progress:
shared._PROGRESS_BAR.completed()
if shared._ERRORS and show_errors:
print('\n%.f Failed download%s:' % (
if shared._ERRORS:
# Send errors to logging module
logger = utils.get_yf_logger()
logger.error('\n%.f Failed download%s:' % (
len(shared._ERRORS), 's' if len(shared._ERRORS) > 1 else ''))
# print(shared._ERRORS)
print("\n".join(['- %s: %s' %
v for v in list(shared._ERRORS.items())]))
# Log each distinct error once, with list of symbols affected
errors = {}
for ticker in shared._ERRORS:
err = shared._ERRORS[ticker]
err = err.replace(f'{ticker}', '%ticker%')
if err not in errors:
errors[err] = [ticker]
else:
errors[err].append(ticker)
for err in errors.keys():
logger.error(f'{errors[err]}: ' + err)
# Log each distinct traceback once, with list of symbols affected
tbs = {}
for ticker in shared._TRACEBACKS:
tb = shared._TRACEBACKS[ticker]
tb = tb.replace(f'{ticker}', '%ticker%')
if tb not in tbs:
tbs[tb] = [ticker]
else:
tbs[tb].append(ticker)
for tb in tbs.keys():
logger.debug(f'{tbs[tb]}: ' + tb)
if ignore_tz:
for tkr in shared._DFS.keys():
@@ -150,16 +213,16 @@ def download(tickers, start=None, end=None, actions=False, threads=True, ignore_
if len(tickers) == 1:
ticker = tickers[0]
return shared._DFS[shared._ISINS.get(ticker, ticker)]
return shared._DFS[ticker]
try:
data = _pd.concat(shared._DFS.values(), axis=1, sort=True,
keys=shared._DFS.keys())
keys=shared._DFS.keys(), names=['Ticker', 'Price'])
except Exception:
_realign_dfs()
data = _pd.concat(shared._DFS.values(), axis=1, sort=True,
keys=shared._DFS.keys())
keys=shared._DFS.keys(), names=['Ticker', 'Price'])
data.index = _pd.to_datetime(data.index)
# switch names back to isins if applicable
data.rename(columns=shared._ISINS, inplace=True)
@@ -199,16 +262,9 @@ def _download_one_threaded(ticker, start=None, end=None,
actions=False, progress=True, period="max",
interval="1d", prepost=False, proxy=None,
keepna=False, rounding=False, timeout=10):
try:
data = _download_one(ticker, start, end, auto_adjust, back_adjust, repair,
actions, period, interval, prepost, proxy, rounding,
keepna, timeout)
except Exception as e:
# glob try/except needed as current thead implementation breaks if exception is raised.
shared._DFS[ticker] = utils.empty_df()
shared._ERRORS[ticker] = repr(e)
else:
shared._DFS[ticker.upper()] = data
_download_one(ticker, start, end, auto_adjust, back_adjust, repair,
actions, period, interval, prepost, proxy, rounding,
keepna, timeout)
if progress:
shared._PROGRESS_BAR.animate()
@@ -218,11 +274,22 @@ def _download_one(ticker, start=None, end=None,
actions=False, period="max", interval="1d",
prepost=False, proxy=None, rounding=False,
keepna=False, timeout=10):
return Ticker(ticker).history(
period=period, interval=interval,
start=start, end=end, prepost=prepost,
actions=actions, auto_adjust=auto_adjust,
back_adjust=back_adjust, repair=repair, proxy=proxy,
rounding=rounding, keepna=keepna, timeout=timeout,
debug=False, raise_errors=False # debug and raise_errors false to not log and raise errors in threads
)
data = None
try:
data = Ticker(ticker).history(
period=period, interval=interval,
start=start, end=end, prepost=prepost,
actions=actions, auto_adjust=auto_adjust,
back_adjust=back_adjust, repair=repair, proxy=proxy,
rounding=rounding, keepna=keepna, timeout=timeout,
raise_errors=True
)
except Exception as e:
# glob try/except needed as current thead implementation breaks if exception is raised.
shared._DFS[ticker.upper()] = utils.empty_df()
shared._ERRORS[ticker.upper()] = repr(e)
shared._TRACEBACKS[ticker.upper()] = traceback.format_exc()
else:
shared._DFS[ticker.upper()] = data
return data

View File

@@ -1,13 +1,14 @@
import pandas as pd
from yfinance import utils
from yfinance.data import TickerData
from yfinance.data import YfData
from yfinance.exceptions import YFNotImplementedError
class Analysis:
def __init__(self, data: TickerData, proxy=None):
def __init__(self, data: YfData, symbol: str, proxy=None):
self._data = data
self._symbol = symbol
self.proxy = proxy
self._earnings_trend = None
@@ -20,99 +21,29 @@ class Analysis:
@property
def earnings_trend(self) -> pd.DataFrame:
if self._earnings_trend is None:
self._scrape(self.proxy)
raise YFNotImplementedError('earnings_trend')
return self._earnings_trend
@property
def analyst_trend_details(self) -> pd.DataFrame:
if self._analyst_trend_details is None:
self._scrape(self.proxy)
raise YFNotImplementedError('analyst_trend_details')
return self._analyst_trend_details
@property
def analyst_price_target(self) -> pd.DataFrame:
if self._analyst_price_target is None:
self._scrape(self.proxy)
raise YFNotImplementedError('analyst_price_target')
return self._analyst_price_target
@property
def rev_est(self) -> pd.DataFrame:
if self._rev_est is None:
self._scrape(self.proxy)
raise YFNotImplementedError('rev_est')
return self._rev_est
@property
def eps_est(self) -> pd.DataFrame:
if self._eps_est is None:
self._scrape(self.proxy)
raise YFNotImplementedError('eps_est')
return self._eps_est
def _scrape(self, proxy):
if self._already_scraped:
return
self._already_scraped = True
# Analysis Data/Analyst Forecasts
analysis_data = self._data.get_json_data_stores("analysis", proxy=proxy)
try:
analysis_data = analysis_data['QuoteSummaryStore']
except KeyError as e:
err_msg = "No analysis data found, symbol may be delisted"
print('- %s: %s' % (self._data.ticker, err_msg))
return
if isinstance(analysis_data.get('earningsTrend'), dict):
try:
analysis = pd.DataFrame(analysis_data['earningsTrend']['trend'])
analysis['endDate'] = pd.to_datetime(analysis['endDate'])
analysis.set_index('period', inplace=True)
analysis.index = analysis.index.str.upper()
analysis.index.name = 'Period'
analysis.columns = utils.camel2title(analysis.columns)
dict_cols = []
for idx, row in analysis.iterrows():
for colname, colval in row.items():
if isinstance(colval, dict):
dict_cols.append(colname)
for k, v in colval.items():
new_colname = colname + ' ' + \
utils.camel2title([k])[0]
analysis.loc[idx, new_colname] = v
self._earnings_trend = analysis[[
c for c in analysis.columns if c not in dict_cols]]
except Exception:
pass
try:
self._analyst_trend_details = pd.DataFrame(analysis_data['recommendationTrend']['trend'])
except Exception as e:
self._analyst_trend_details = None
try:
self._analyst_price_target = pd.DataFrame(analysis_data['financialData'], index=[0])[
['targetLowPrice', 'currentPrice', 'targetMeanPrice', 'targetHighPrice', 'numberOfAnalystOpinions']].T
except Exception as e:
self._analyst_price_target = None
earnings_estimate = []
revenue_estimate = []
if self._analyst_trend_details is not None :
for key in analysis_data['earningsTrend']['trend']:
try:
earnings_dict = key['earningsEstimate']
earnings_dict['period'] = key['period']
earnings_dict['endDate'] = key['endDate']
earnings_estimate.append(earnings_dict)
revenue_dict = key['revenueEstimate']
revenue_dict['period'] = key['period']
revenue_dict['endDate'] = key['endDate']
revenue_estimate.append(revenue_dict)
except Exception as e:
pass
self._rev_est = pd.DataFrame(revenue_estimate)
self._eps_est = pd.DataFrame(earnings_estimate)
else:
self._rev_est = pd.DataFrame()
self._eps_est = pd.DataFrame()

View File

@@ -2,17 +2,17 @@ import datetime
import json
import pandas as pd
import numpy as np
from yfinance import utils
from yfinance.data import TickerData
from yfinance.exceptions import YFinanceDataException, YFinanceException
from yfinance import utils, const
from yfinance.data import YfData
from yfinance.exceptions import YFinanceException, YFNotImplementedError
class Fundamentals:
def __init__(self, data: TickerData, proxy=None):
def __init__(self, data: YfData, symbol: str, proxy=None):
self._data = data
self._symbol = symbol
self.proxy = proxy
self._earnings = None
@@ -22,7 +22,7 @@ class Fundamentals:
self._financials_data = None
self._fin_data_quote = None
self._basics_already_scraped = False
self._financials = Financials(data)
self._financials = Financials(data, symbol)
@property
def financials(self) -> "Financials":
@@ -31,100 +31,43 @@ class Fundamentals:
@property
def earnings(self) -> dict:
if self._earnings is None:
self._scrape_earnings(self.proxy)
raise YFNotImplementedError('earnings')
return self._earnings
@property
def shares(self) -> pd.DataFrame:
if self._shares is None:
self._scrape_shares(self.proxy)
raise YFNotImplementedError('shares')
return self._shares
def _scrape_basics(self, proxy):
if self._basics_already_scraped:
return
self._basics_already_scraped = True
self._financials_data = self._data.get_json_data_stores('financials', proxy)
try:
self._fin_data_quote = self._financials_data['QuoteSummaryStore']
except KeyError:
err_msg = "No financials data found, symbol may be delisted"
print('- %s: %s' % (self._data.ticker, err_msg))
return None
def _scrape_earnings(self, proxy):
self._scrape_basics(proxy)
# earnings
self._earnings = {"yearly": pd.DataFrame(), "quarterly": pd.DataFrame()}
if self._fin_data_quote is None:
return
if isinstance(self._fin_data_quote.get('earnings'), dict):
try:
earnings = self._fin_data_quote['earnings']['financialsChart']
earnings['financialCurrency'] = self._fin_data_quote['earnings'].get('financialCurrency', 'USD')
self._earnings['financialCurrency'] = earnings['financialCurrency']
df = pd.DataFrame(earnings['yearly']).set_index('date')
df.columns = utils.camel2title(df.columns)
df.index.name = 'Year'
self._earnings['yearly'] = df
df = pd.DataFrame(earnings['quarterly']).set_index('date')
df.columns = utils.camel2title(df.columns)
df.index.name = 'Quarter'
self._earnings['quarterly'] = df
except Exception:
pass
def _scrape_shares(self, proxy):
self._scrape_basics(proxy)
# shares outstanding
try:
# keep only years with non None data
available_shares = [shares_data for shares_data in
self._financials_data['QuoteTimeSeriesStore']['timeSeries']['annualBasicAverageShares']
if
shares_data]
shares = pd.DataFrame(available_shares)
shares['Year'] = shares['asOfDate'].agg(lambda x: int(x[:4]))
shares.set_index('Year', inplace=True)
shares.drop(columns=['dataId', 'asOfDate',
'periodType', 'currencyCode'], inplace=True)
shares.rename(
columns={'reportedValue': "BasicShares"}, inplace=True)
self._shares = shares
except Exception:
pass
class Financials:
def __init__(self, data: TickerData):
def __init__(self, data: YfData, symbol: str):
self._data = data
self._symbol = symbol
self._income_time_series = {}
self._balance_sheet_time_series = {}
self._cash_flow_time_series = {}
self._income_scraped = {}
self._balance_sheet_scraped = {}
self._cash_flow_scraped = {}
def get_income_time_series(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._income_time_series
if freq not in res:
res[freq] = self._fetch_time_series("income", freq, proxy=None)
res[freq] = self._fetch_time_series("income", freq, proxy)
return res[freq]
def get_balance_sheet_time_series(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._balance_sheet_time_series
if freq not in res:
res[freq] = self._fetch_time_series("balance-sheet", freq, proxy=None)
res[freq] = self._fetch_time_series("balance-sheet", freq, proxy)
return res[freq]
def get_cash_flow_time_series(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._cash_flow_time_series
if freq not in res:
res[freq] = self._fetch_time_series("cash-flow", freq, proxy=None)
res[freq] = self._fetch_time_series("cash-flow", freq, proxy)
return res[freq]
@utils.log_indent_decorator
def _fetch_time_series(self, name, timescale, proxy=None):
# Fetching time series preferred over scraping 'QuoteSummaryStore',
# because it matches what Yahoo shows. But for some tickers returns nothing,
@@ -134,9 +77,9 @@ class Financials:
allowed_timescales = ["yearly", "quarterly"]
if name not in allowed_names:
raise ValueError("Illegal argument: name must be one of: {}".format(allowed_names))
raise ValueError(f"Illegal argument: name must be one of: {allowed_names}")
if timescale not in allowed_timescales:
raise ValueError("Illegal argument: timescale must be one of: {}".format(allowed_names))
raise ValueError(f"Illegal argument: timescale must be one of: {allowed_timescales}")
try:
statement = self._create_financials_table(name, timescale, proxy)
@@ -144,7 +87,7 @@ class Financials:
if statement is not None:
return statement
except YFinanceException as e:
print(f"- {self._data.ticker}: Failed to create {name} financials table for reason: {repr(e)}")
utils.get_yf_logger().error(f"{self._symbol}: Failed to create {name} financials table for reason: {e}")
return pd.DataFrame()
def _create_financials_table(self, name, timescale, proxy):
@@ -152,51 +95,24 @@ class Financials:
# Yahoo stores the 'income' table internally under 'financials' key
name = "financials"
keys = self._get_datastore_keys(name, proxy)
keys = const.fundamentals_keys[name]
try:
return self.get_financials_time_series(timescale, keys, proxy)
except Exception as e:
except Exception:
pass
def _get_datastore_keys(self, sub_page, proxy) -> list:
data_stores = self._data.get_json_data_stores(sub_page, proxy)
# Step 1: get the keys:
def _finditem1(key, obj):
values = []
if isinstance(obj, dict):
if key in obj.keys():
values.append(obj[key])
for k, v in obj.items():
values += _finditem1(key, v)
elif isinstance(obj, list):
for v in obj:
values += _finditem1(key, v)
return values
try:
keys = _finditem1("key", data_stores['FinancialTemplateStore'])
except KeyError as e:
raise YFinanceDataException("Parsing FinancialTemplateStore failed, reason: {}".format(repr(e)))
if not keys:
raise YFinanceDataException("No keys in FinancialTemplateStore")
return keys
def get_financials_time_series(self, timescale, keys: list, proxy=None) -> pd.DataFrame:
timescale_translation = {"yearly": "annual", "quarterly": "quarterly"}
timescale = timescale_translation[timescale]
# Step 2: construct url:
ts_url_base = \
"https://query2.finance.yahoo.com/ws/fundamentals-timeseries/v1/finance/timeseries/{0}?symbol={0}" \
.format(self._data.ticker)
ts_url_base = f"https://query2.finance.yahoo.com/ws/fundamentals-timeseries/v1/finance/timeseries/{self._symbol}?symbol={self._symbol}"
url = ts_url_base + "&type=" + ",".join([timescale + k for k in keys])
# Yahoo returns maximum 4 years or 5 quarters, regardless of start_dt:
start_dt = datetime.datetime(2016, 12, 31)
end = pd.Timestamp.utcnow().ceil("D")
url += "&period1={}&period2={}".format(int(start_dt.timestamp()), int(end.timestamp()))
url += f"&period1={int(start_dt.timestamp())}&period2={int(end.timestamp())}"
# Step 3: fetch and reshape data
json_str = self._data.cache_get(url=url, proxy=proxy).text
@@ -231,89 +147,3 @@ class Financials:
df = df[sorted(df.columns, reverse=True)]
return df
def get_income_scrape(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._income_scraped
if freq not in res:
res[freq] = self._scrape("income", freq, proxy=None)
return res[freq]
def get_balance_sheet_scrape(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._balance_sheet_scraped
if freq not in res:
res[freq] = self._scrape("balance-sheet", freq, proxy=None)
return res[freq]
def get_cash_flow_scrape(self, freq="yearly", proxy=None) -> pd.DataFrame:
res = self._cash_flow_scraped
if freq not in res:
res[freq] = self._scrape("cash-flow", freq, proxy=None)
return res[freq]
def _scrape(self, name, timescale, proxy=None):
# Backup in case _fetch_time_series() fails to return data
allowed_names = ["income", "balance-sheet", "cash-flow"]
allowed_timescales = ["yearly", "quarterly"]
if name not in allowed_names:
raise ValueError("Illegal argument: name must be one of: {}".format(allowed_names))
if timescale not in allowed_timescales:
raise ValueError("Illegal argument: timescale must be one of: {}".format(allowed_names))
try:
statement = self._create_financials_table_old(name, timescale, proxy)
if statement is not None:
return statement
except YFinanceException as e:
print(f"- {self._data.ticker}: Failed to create financials table for {name} reason: {repr(e)}")
return pd.DataFrame()
def _create_financials_table_old(self, name, timescale, proxy):
data_stores = self._data.get_json_data_stores("financials", proxy)
# Fetch raw data
if not "QuoteSummaryStore" in data_stores:
raise YFinanceDataException(f"Yahoo not returning legacy financials data")
data = data_stores["QuoteSummaryStore"]
if name == "cash-flow":
key1 = "cashflowStatement"
key2 = "cashflowStatements"
elif name == "balance-sheet":
key1 = "balanceSheet"
key2 = "balanceSheetStatements"
else:
key1 = "incomeStatement"
key2 = "incomeStatementHistory"
key1 += "History"
if timescale == "quarterly":
key1 += "Quarterly"
if key1 not in data or data[key1] is None or key2 not in data[key1]:
raise YFinanceDataException(f"Yahoo not returning legacy {name} financials data")
data = data[key1][key2]
# Tabulate
df = pd.DataFrame(data)
if len(df) == 0:
raise YFinanceDataException(f"Yahoo not returning legacy {name} financials data")
df = df.drop(columns=['maxAge'])
for col in df.columns:
df[col] = df[col].replace('-', np.nan)
df.set_index('endDate', inplace=True)
try:
df.index = pd.to_datetime(df.index, unit='s')
except ValueError:
df.index = pd.to_datetime(df.index)
df = df.T
df.columns.name = ''
df.index.name = 'Breakdown'
# rename incorrect yahoo key
df.rename(index={'treasuryStock': 'gainsLossesNotAffectingRetainedEarnings'}, inplace=True)
# Upper-case first letter, leave rest unchanged:
s0 = df.index[0]
df.index = [s[0].upper()+s[1:] for s in df.index]
return df

1630
yfinance/scrapers/history.py Normal file

File diff suppressed because it is too large Load Diff

View File

@@ -1,66 +1,246 @@
import pandas as pd
# from io import StringIO
import pandas as pd
import requests
from yfinance import utils
from yfinance.data import YfData
from yfinance.const import _BASE_URL_
from yfinance.exceptions import YFinanceDataException
_QUOTE_SUMMARY_URL_ = f"{_BASE_URL_}/v10/finance/quoteSummary/"
from yfinance.data import TickerData
class Holders:
_SCRAPE_URL_ = 'https://finance.yahoo.com/quote'
def __init__(self, data: TickerData, proxy=None):
def __init__(self, data: YfData, symbol: str, proxy=None):
self._data = data
self._symbol = symbol
self.proxy = proxy
self._major = None
self._major_direct_holders = None
self._institutional = None
self._mutualfund = None
self._insider_transactions = None
self._insider_purchases = None
self._insider_roster = None
@property
def major(self) -> pd.DataFrame:
if self._major is None:
self._scrape(self.proxy)
# self._scrape(self.proxy)
self._fetch_and_parse()
return self._major
@property
def institutional(self) -> pd.DataFrame:
if self._institutional is None:
self._scrape(self.proxy)
# self._scrape(self.proxy)
self._fetch_and_parse()
return self._institutional
@property
def mutualfund(self) -> pd.DataFrame:
if self._mutualfund is None:
self._scrape(self.proxy)
# self._scrape(self.proxy)
self._fetch_and_parse()
return self._mutualfund
def _scrape(self, proxy):
ticker_url = "{}/{}".format(self._SCRAPE_URL_, self._data.ticker)
@property
def insider_transactions(self) -> pd.DataFrame:
if self._insider_transactions is None:
# self._scrape_insider_transactions(self.proxy)
self._fetch_and_parse()
return self._insider_transactions
@property
def insider_purchases(self) -> pd.DataFrame:
if self._insider_purchases is None:
# self._scrape_insider_transactions(self.proxy)
self._fetch_and_parse()
return self._insider_purchases
@property
def insider_roster(self) -> pd.DataFrame:
if self._insider_roster is None:
# self._scrape_insider_ros(self.proxy)
self._fetch_and_parse()
return self._insider_roster
def _fetch(self, proxy):
modules = ','.join(
["institutionOwnership", "fundOwnership", "majorDirectHolders", "majorHoldersBreakdown", "insiderTransactions", "insiderHolders", "netSharePurchaseActivity"])
params_dict = {"modules": modules, "corsDomain": "finance.yahoo.com", "formatted": "false"}
result = self._data.get_raw_json(f"{_QUOTE_SUMMARY_URL_}/{self._symbol}", user_agent_headers=self._data.user_agent_headers, params=params_dict, proxy=proxy)
return result
def _fetch_and_parse(self):
try:
resp = self._data.cache_get(ticker_url + '/holders', proxy)
holders = pd.read_html(resp.text)
except Exception:
holders = []
result = self._fetch(self.proxy)
except requests.exceptions.HTTPError as e:
utils.get_yf_logger().error(str(e))
if len(holders) >= 3:
self._major = holders[0]
self._institutional = holders[1]
self._mutualfund = holders[2]
elif len(holders) >= 2:
self._major = holders[0]
self._institutional = holders[1]
elif len(holders) >= 1:
self._major = holders[0]
self._major = pd.DataFrame()
self._major_direct_holders = pd.DataFrame()
self._institutional = pd.DataFrame()
self._mutualfund = pd.DataFrame()
self._insider_transactions = pd.DataFrame()
self._insider_purchases = pd.DataFrame()
self._insider_roster = pd.DataFrame()
if self._institutional is not None:
if 'Date Reported' in self._institutional:
self._institutional['Date Reported'] = pd.to_datetime(
self._institutional['Date Reported'])
if '% Out' in self._institutional:
self._institutional['% Out'] = self._institutional[
'% Out'].str.replace('%', '').astype(float) / 100
return
if self._mutualfund is not None:
if 'Date Reported' in self._mutualfund:
self._mutualfund['Date Reported'] = pd.to_datetime(
self._mutualfund['Date Reported'])
if '% Out' in self._mutualfund:
self._mutualfund['% Out'] = self._mutualfund[
'% Out'].str.replace('%', '').astype(float) / 100
try:
data = result["quoteSummary"]["result"][0]
# parse "institutionOwnership", "fundOwnership", "majorDirectHolders", "majorHoldersBreakdown", "insiderTransactions", "insiderHolders", "netSharePurchaseActivity"
self._parse_institution_ownership(data["institutionOwnership"])
self._parse_fund_ownership(data["fundOwnership"])
# self._parse_major_direct_holders(data["majorDirectHolders"]) # need more data to investigate
self._parse_major_holders_breakdown(data["majorHoldersBreakdown"])
self._parse_insider_transactions(data["insiderTransactions"])
self._parse_insider_holders(data["insiderHolders"])
self._parse_net_share_purchase_activity(data["netSharePurchaseActivity"])
except (KeyError, IndexError):
raise YFinanceDataException("Failed to parse holders json data.")
@staticmethod
def _parse_raw_values(data):
if isinstance(data, dict) and "raw" in data:
return data["raw"]
return data
def _parse_institution_ownership(self, data):
holders = data["ownershipList"]
for owner in holders:
for k, v in owner.items():
owner[k] = self._parse_raw_values(v)
del owner["maxAge"]
df = pd.DataFrame(holders)
if not df.empty:
df["reportDate"] = pd.to_datetime(df["reportDate"], unit="s")
df.rename(columns={"reportDate": "Date Reported", "organization": "Holder", "position": "Shares", "value": "Value"}, inplace=True) # "pctHeld": "% Out"
self._institutional = df
def _parse_fund_ownership(self, data):
holders = data["ownershipList"]
for owner in holders:
for k, v in owner.items():
owner[k] = self._parse_raw_values(v)
del owner["maxAge"]
df = pd.DataFrame(holders)
if not df.empty:
df["reportDate"] = pd.to_datetime(df["reportDate"], unit="s")
df.rename(columns={"reportDate": "Date Reported", "organization": "Holder", "position": "Shares", "value": "Value"}, inplace=True)
self._mutualfund = df
def _parse_major_direct_holders(self, data):
holders = data["holders"]
for owner in holders:
for k, v in owner.items():
owner[k] = self._parse_raw_values(v)
del owner["maxAge"]
df = pd.DataFrame(holders)
if not df.empty:
df["reportDate"] = pd.to_datetime(df["reportDate"], unit="s")
df.rename(columns={"reportDate": "Date Reported", "organization": "Holder", "positionDirect": "Shares", "valueDirect": "Value"}, inplace=True)
self._major_direct_holders = df
def _parse_major_holders_breakdown(self, data):
if "maxAge" in data:
del data["maxAge"]
df = pd.DataFrame.from_dict(data, orient="index")
if not df.empty:
df.columns.name = "Breakdown"
df.rename(columns={df.columns[0]: 'Value'}, inplace=True)
self._major = df
def _parse_insider_transactions(self, data):
holders = data["transactions"]
for owner in holders:
for k, v in owner.items():
owner[k] = self._parse_raw_values(v)
del owner["maxAge"]
df = pd.DataFrame(holders)
if not df.empty:
df["startDate"] = pd.to_datetime(df["startDate"], unit="s")
df.rename(columns={
"startDate": "Start Date",
"filerName": "Insider",
"filerRelation": "Position",
"filerUrl": "URL",
"moneyText": "Transaction",
"transactionText": "Text",
"shares": "Shares",
"value": "Value",
"ownership": "Ownership" # ownership flag, direct or institutional
}, inplace=True)
self._insider_transactions = df
def _parse_insider_holders(self, data):
holders = data["holders"]
for owner in holders:
for k, v in owner.items():
owner[k] = self._parse_raw_values(v)
del owner["maxAge"]
df = pd.DataFrame(holders)
if not df.empty:
df["positionDirectDate"] = pd.to_datetime(df["positionDirectDate"], unit="s")
df["latestTransDate"] = pd.to_datetime(df["latestTransDate"], unit="s")
df.rename(columns={
"name": "Name",
"relation": "Position",
"url": "URL",
"transactionDescription": "Most Recent Transaction",
"latestTransDate": "Latest Transaction Date",
"positionDirectDate": "Position Direct Date",
"positionDirect": "Shares Owned Directly",
"positionIndirectDate": "Position Indirect Date",
"positionIndirect": "Shares Owned Indirectly"
}, inplace=True)
df["Name"] = df["Name"].astype(str)
df["Position"] = df["Position"].astype(str)
df["URL"] = df["URL"].astype(str)
df["Most Recent Transaction"] = df["Most Recent Transaction"].astype(str)
self._insider_roster = df
def _parse_net_share_purchase_activity(self, data):
df = pd.DataFrame(
{
"Insider Purchases Last " + data.get("period", ""): [
"Purchases",
"Sales",
"Net Shares Purchased (Sold)",
"Total Insider Shares Held",
"% Net Shares Purchased (Sold)",
"% Buy Shares",
"% Sell Shares"
],
"Shares": [
data.get('buyInfoShares'),
data.get('sellInfoShares'),
data.get('netInfoShares'),
data.get('totalInsiderShares'),
data.get('netPercentInsiderShares'),
data.get('buyPercentInsiderShares'),
data.get('sellPercentInsiderShares')
],
"Trans": [
data.get('buyInfoCount'),
data.get('sellInfoCount'),
data.get('netInfoCount'),
pd.NA,
pd.NA,
pd.NA,
pd.NA
]
}
).convert_dtypes()
self._insider_purchases = df

View File

@@ -1,31 +1,32 @@
import datetime
import json
import warnings
from collections.abc import MutableMapping
import numpy as _np
import pandas as pd
import requests
from yfinance import utils
from yfinance.data import TickerData
from yfinance.data import YfData
from yfinance.const import quote_summary_valid_modules, _BASE_URL_
from yfinance.exceptions import YFNotImplementedError, YFinanceDataException, YFinanceException
info_retired_keys_price = {"currentPrice", "dayHigh", "dayLow", "open", "previousClose", "volume"}
info_retired_keys_price = {"currentPrice", "dayHigh", "dayLow", "open", "previousClose", "volume", "volume24Hr"}
info_retired_keys_price.update({"regularMarket"+s for s in ["DayHigh", "DayLow", "Open", "PreviousClose", "Price", "Volume"]})
info_retired_keys_price.update({"fiftyTwoWeekLow", "fiftyTwoWeekHigh", "fiftyTwoWeekChange", "fiftyDayAverage", "twoHundredDayAverage"})
info_retired_keys_price.update({"fiftyTwoWeekLow", "fiftyTwoWeekHigh", "fiftyTwoWeekChange", "52WeekChange", "fiftyDayAverage", "twoHundredDayAverage"})
info_retired_keys_price.update({"averageDailyVolume10Day", "averageVolume10days", "averageVolume"})
info_retired_keys_exchange = {"currency", "exchange", "exchangeTimezoneName", "exchangeTimezoneShortName"}
info_retired_keys_exchange = {"currency", "exchange", "exchangeTimezoneName", "exchangeTimezoneShortName", "quoteType"}
info_retired_keys_marketCap = {"marketCap"}
info_retired_keys_symbol = {"symbol"}
info_retired_keys = info_retired_keys_price | info_retired_keys_exchange | info_retired_keys_marketCap | info_retired_keys_symbol
#
info_retired_keys = []
PRUNE_INFO = True
# PRUNE_INFO = False
_QUOTE_SUMMARY_URL_ = f"{_BASE_URL_}/v10/finance/quoteSummary"
from collections.abc import MutableMapping
class InfoDictWrapper(MutableMapping):
""" Simple wrapper around info dict, intercepting 'gets' to
""" Simple wrapper around info dict, intercepting 'gets' to
print how-to-migrate messages for specific keys. Requires
override dict API"""
@@ -46,16 +47,16 @@ class InfoDictWrapper(MutableMapping):
def __getitem__(self, k):
if k in info_retired_keys_price:
print(f"Price data removed from info. Use Ticker.basic_info or history() instead")
warnings.warn(f"Price data removed from info (key='{k}'). Use Ticker.fast_info or history() instead", DeprecationWarning)
return None
elif k in info_retired_keys_exchange:
print(f"Exchange data removed from info. Use Ticker.basic_info or Ticker.get_history_metadata() instead")
warnings.warn(f"Exchange data removed from info (key='{k}'). Use Ticker.fast_info or Ticker.get_history_metadata() instead", DeprecationWarning)
return None
elif k in info_retired_keys_marketCap:
print(f"Market cap removed from info. Use Ticker.basic_info instead")
warnings.warn(f"Market cap removed from info (key='{k}'). Use Ticker.fast_info instead", DeprecationWarning)
return None
elif k in info_retired_keys_symbol:
print(f"Symbol removed from info. You know this already")
warnings.warn(f"Symbol removed from info (key='{k}'). You know this already", DeprecationWarning)
return None
return self.info[self._keytransform(k)]
@@ -67,7 +68,7 @@ class InfoDictWrapper(MutableMapping):
def __iter__(self):
return iter(self.info)
def __len__(self):
return len(self.info)
@@ -75,170 +76,609 @@ class InfoDictWrapper(MutableMapping):
return k
class FastInfo:
# Contain small subset of info[] items that can be fetched faster elsewhere.
# Imitates a dict.
def __init__(self, tickerBaseObject, proxy=None):
self._tkr = tickerBaseObject
self.proxy = proxy
self._prices_1y = None
self._prices_1wk_1h_prepost = None
self._prices_1wk_1h_reg = None
self._md = None
self._currency = None
self._quote_type = None
self._exchange = None
self._timezone = None
self._shares = None
self._mcap = None
self._open = None
self._day_high = None
self._day_low = None
self._last_price = None
self._last_volume = None
self._prev_close = None
self._reg_prev_close = None
self._50d_day_average = None
self._200d_day_average = None
self._year_high = None
self._year_low = None
self._year_change = None
self._10d_avg_vol = None
self._3mo_avg_vol = None
# attrs = utils.attributes(self)
# self.keys = attrs.keys()
# utils.attributes is calling each method, bad! Have to hardcode
_properties = ["currency", "quote_type", "exchange", "timezone"]
_properties += ["shares", "market_cap"]
_properties += ["last_price", "previous_close", "open", "day_high", "day_low"]
_properties += ["regular_market_previous_close"]
_properties += ["last_volume"]
_properties += ["fifty_day_average", "two_hundred_day_average", "ten_day_average_volume", "three_month_average_volume"]
_properties += ["year_high", "year_low", "year_change"]
# Because released before fixing key case, need to officially support
# camel-case but also secretly support snake-case
base_keys = [k for k in _properties if '_' not in k]
sc_keys = [k for k in _properties if '_' in k]
self._sc_to_cc_key = {k: utils.snake_case_2_camelCase(k) for k in sc_keys}
self._cc_to_sc_key = {v: k for k, v in self._sc_to_cc_key.items()}
self._public_keys = sorted(base_keys + list(self._sc_to_cc_key.values()))
self._keys = sorted(self._public_keys + sc_keys)
# dict imitation:
def keys(self):
return self._public_keys
def items(self):
return [(k, self[k]) for k in self._public_keys]
def values(self):
return [self[k] for k in self._public_keys]
def get(self, key, default=None):
if key in self.keys():
if key in self._cc_to_sc_key:
key = self._cc_to_sc_key[key]
return self[key]
return default
def __getitem__(self, k):
if not isinstance(k, str):
raise KeyError("key must be a string")
if k not in self._keys:
raise KeyError(f"'{k}' not valid key. Examine 'FastInfo.keys()'")
if k in self._cc_to_sc_key:
k = self._cc_to_sc_key[k]
return getattr(self, k)
def __contains__(self, k):
return k in self.keys()
def __iter__(self):
return iter(self.keys())
def __str__(self):
return "lazy-loading dict with keys = " + str(self.keys())
def __repr__(self):
return self.__str__()
def toJSON(self, indent=4):
return json.dumps({k: self[k] for k in self.keys()}, indent=indent)
def _get_1y_prices(self, fullDaysOnly=False):
if self._prices_1y is None:
self._prices_1y = self._tkr.history(period="380d", auto_adjust=False, keepna=True, proxy=self.proxy)
self._md = self._tkr.get_history_metadata(proxy=self.proxy)
try:
ctp = self._md["currentTradingPeriod"]
self._today_open = pd.to_datetime(ctp["regular"]["start"], unit='s', utc=True).tz_convert(self.timezone)
self._today_close = pd.to_datetime(ctp["regular"]["end"], unit='s', utc=True).tz_convert(self.timezone)
self._today_midnight = self._today_close.ceil("D")
except Exception:
self._today_open = None
self._today_close = None
self._today_midnight = None
raise
if self._prices_1y.empty:
return self._prices_1y
dnow = pd.Timestamp.utcnow().tz_convert(self.timezone).date()
d1 = dnow
d0 = (d1 + datetime.timedelta(days=1)) - utils._interval_to_timedelta("1y")
if fullDaysOnly and self._exchange_open_now():
# Exclude today
d1 -= utils._interval_to_timedelta("1d")
return self._prices_1y.loc[str(d0):str(d1)]
def _get_1wk_1h_prepost_prices(self):
if self._prices_1wk_1h_prepost is None:
self._prices_1wk_1h_prepost = self._tkr.history(period="1wk", interval="1h", auto_adjust=False, prepost=True, proxy=self.proxy)
return self._prices_1wk_1h_prepost
def _get_1wk_1h_reg_prices(self):
if self._prices_1wk_1h_reg is None:
self._prices_1wk_1h_reg = self._tkr.history(period="1wk", interval="1h", auto_adjust=False, prepost=False, proxy=self.proxy)
return self._prices_1wk_1h_reg
def _get_exchange_metadata(self):
if self._md is not None:
return self._md
self._get_1y_prices()
self._md = self._tkr.get_history_metadata(proxy=self.proxy)
return self._md
def _exchange_open_now(self):
t = pd.Timestamp.utcnow()
self._get_exchange_metadata()
# if self._today_open is None and self._today_close is None:
# r = False
# else:
# r = self._today_open <= t and t < self._today_close
# if self._today_midnight is None:
# r = False
# elif self._today_midnight.date() > t.tz_convert(self.timezone).date():
# r = False
# else:
# r = t < self._today_midnight
last_day_cutoff = self._get_1y_prices().index[-1] + datetime.timedelta(days=1)
last_day_cutoff += datetime.timedelta(minutes=20)
r = t < last_day_cutoff
# print("_exchange_open_now() returning", r)
return r
@property
def currency(self):
if self._currency is not None:
return self._currency
md = self._tkr.get_history_metadata(proxy=self.proxy)
self._currency = md["currency"]
return self._currency
@property
def quote_type(self):
if self._quote_type is not None:
return self._quote_type
md = self._tkr.get_history_metadata(proxy=self.proxy)
self._quote_type = md["instrumentType"]
return self._quote_type
@property
def exchange(self):
if self._exchange is not None:
return self._exchange
self._exchange = self._get_exchange_metadata()["exchangeName"]
return self._exchange
@property
def timezone(self):
if self._timezone is not None:
return self._timezone
self._timezone = self._get_exchange_metadata()["exchangeTimezoneName"]
return self._timezone
@property
def shares(self):
if self._shares is not None:
return self._shares
shares = self._tkr.get_shares_full(start=pd.Timestamp.utcnow().date()-pd.Timedelta(days=548), proxy=self.proxy)
# if shares is None:
# # Requesting 18 months failed, so fallback to shares which should include last year
# shares = self._tkr.get_shares()
if shares is not None:
if isinstance(shares, pd.DataFrame):
shares = shares[shares.columns[0]]
self._shares = int(shares.iloc[-1])
return self._shares
@property
def last_price(self):
if self._last_price is not None:
return self._last_price
prices = self._get_1y_prices()
if prices.empty:
md = self._get_exchange_metadata()
if "regularMarketPrice" in md:
self._last_price = md["regularMarketPrice"]
else:
self._last_price = float(prices["Close"].iloc[-1])
if _np.isnan(self._last_price):
md = self._get_exchange_metadata()
if "regularMarketPrice" in md:
self._last_price = md["regularMarketPrice"]
return self._last_price
@property
def previous_close(self):
if self._prev_close is not None:
return self._prev_close
prices = self._get_1wk_1h_prepost_prices()
fail = False
if prices.empty:
fail = True
else:
prices = prices[["Close"]].groupby(prices.index.date).last()
if prices.shape[0] < 2:
# Very few symbols have previousClose despite no
# no trading data e.g. 'QCSTIX'.
fail = True
else:
self._prev_close = float(prices["Close"].iloc[-2])
if fail:
# Fallback to original info[] if available.
self._tkr.info # trigger fetch
k = "previousClose"
if self._tkr._quote._retired_info is not None and k in self._tkr._quote._retired_info:
self._prev_close = self._tkr._quote._retired_info[k]
return self._prev_close
@property
def regular_market_previous_close(self):
if self._reg_prev_close is not None:
return self._reg_prev_close
prices = self._get_1y_prices()
if prices.shape[0] == 1:
# Tiny % of tickers don't return daily history before last trading day,
# so backup option is hourly history:
prices = self._get_1wk_1h_reg_prices()
prices = prices[["Close"]].groupby(prices.index.date).last()
if prices.shape[0] < 2:
# Very few symbols have regularMarketPreviousClose despite no
# no trading data. E.g. 'QCSTIX'.
# So fallback to original info[] if available.
self._tkr.info # trigger fetch
k = "regularMarketPreviousClose"
if self._tkr._quote._retired_info is not None and k in self._tkr._quote._retired_info:
self._reg_prev_close = self._tkr._quote._retired_info[k]
else:
self._reg_prev_close = float(prices["Close"].iloc[-2])
return self._reg_prev_close
@property
def open(self):
if self._open is not None:
return self._open
prices = self._get_1y_prices()
if prices.empty:
self._open = None
else:
self._open = float(prices["Open"].iloc[-1])
if _np.isnan(self._open):
self._open = None
return self._open
@property
def day_high(self):
if self._day_high is not None:
return self._day_high
prices = self._get_1y_prices()
if prices.empty:
self._day_high = None
else:
self._day_high = float(prices["High"].iloc[-1])
if _np.isnan(self._day_high):
self._day_high = None
return self._day_high
@property
def day_low(self):
if self._day_low is not None:
return self._day_low
prices = self._get_1y_prices()
if prices.empty:
self._day_low = None
else:
self._day_low = float(prices["Low"].iloc[-1])
if _np.isnan(self._day_low):
self._day_low = None
return self._day_low
@property
def last_volume(self):
if self._last_volume is not None:
return self._last_volume
prices = self._get_1y_prices()
self._last_volume = None if prices.empty else int(prices["Volume"].iloc[-1])
return self._last_volume
@property
def fifty_day_average(self):
if self._50d_day_average is not None:
return self._50d_day_average
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
self._50d_day_average = None
else:
n = prices.shape[0]
a = n-50
b = n
if a < 0:
a = 0
self._50d_day_average = float(prices["Close"].iloc[a:b].mean())
return self._50d_day_average
@property
def two_hundred_day_average(self):
if self._200d_day_average is not None:
return self._200d_day_average
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
self._200d_day_average = None
else:
n = prices.shape[0]
a = n-200
b = n
if a < 0:
a = 0
self._200d_day_average = float(prices["Close"].iloc[a:b].mean())
return self._200d_day_average
@property
def ten_day_average_volume(self):
if self._10d_avg_vol is not None:
return self._10d_avg_vol
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
self._10d_avg_vol = None
else:
n = prices.shape[0]
a = n-10
b = n
if a < 0:
a = 0
self._10d_avg_vol = int(prices["Volume"].iloc[a:b].mean())
return self._10d_avg_vol
@property
def three_month_average_volume(self):
if self._3mo_avg_vol is not None:
return self._3mo_avg_vol
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
self._3mo_avg_vol = None
else:
dt1 = prices.index[-1]
dt0 = dt1 - utils._interval_to_timedelta("3mo") + utils._interval_to_timedelta("1d")
self._3mo_avg_vol = int(prices.loc[dt0:dt1, "Volume"].mean())
return self._3mo_avg_vol
@property
def year_high(self):
if self._year_high is not None:
return self._year_high
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
prices = self._get_1y_prices(fullDaysOnly=False)
self._year_high = float(prices["High"].max())
return self._year_high
@property
def year_low(self):
if self._year_low is not None:
return self._year_low
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.empty:
prices = self._get_1y_prices(fullDaysOnly=False)
self._year_low = float(prices["Low"].min())
return self._year_low
@property
def year_change(self):
if self._year_change is not None:
return self._year_change
prices = self._get_1y_prices(fullDaysOnly=True)
if prices.shape[0] >= 2:
self._year_change = (prices["Close"].iloc[-1] - prices["Close"].iloc[0]) / prices["Close"].iloc[0]
self._year_change = float(self._year_change)
return self._year_change
@property
def market_cap(self):
if self._mcap is not None:
return self._mcap
try:
shares = self.shares
except Exception as e:
if "Cannot retrieve share count" in str(e):
shares = None
elif "failed to decrypt Yahoo" in str(e):
shares = None
else:
raise
if shares is None:
# Very few symbols have marketCap despite no share count.
# E.g. 'BTC-USD'
# So fallback to original info[] if available.
self._tkr.info
k = "marketCap"
if self._tkr._quote._retired_info is not None and k in self._tkr._quote._retired_info:
self._mcap = self._tkr._quote._retired_info[k]
else:
self._mcap = float(shares * self.last_price)
return self._mcap
class Quote:
def __init__(self, data: TickerData, proxy=None):
def __init__(self, data: YfData, symbol: str, proxy=None):
self._data = data
self._symbol = symbol
self.proxy = proxy
self._info = None
self._retired_info = None
self._sustainability = None
self._recommendations = None
self._upgrades_downgrades = None
self._calendar = None
self._already_scraped = False
self._already_scraped_complementary = False
self._already_fetched = False
self._already_fetched_complementary = False
@property
def info(self) -> dict:
if self._info is None:
self._scrape(self.proxy)
self._scrape_complementary(self.proxy)
self._fetch_info(self.proxy)
self._fetch_complementary(self.proxy)
return self._info
@property
def sustainability(self) -> pd.DataFrame:
if self._sustainability is None:
self._scrape(self.proxy)
raise YFNotImplementedError('sustainability')
return self._sustainability
@property
def recommendations(self) -> pd.DataFrame:
if self._recommendations is None:
self._scrape(self.proxy)
result = self._fetch(self.proxy, modules=['recommendationTrend'])
if result is None:
self._recommendations = pd.DataFrame()
else:
try:
data = result["quoteSummary"]["result"][0]["recommendationTrend"]["trend"]
except (KeyError, IndexError):
raise YFinanceDataException(f"Failed to parse json response from Yahoo Finance: {result}")
self._recommendations = pd.DataFrame(data)
return self._recommendations
@property
def calendar(self) -> pd.DataFrame:
def upgrades_downgrades(self) -> pd.DataFrame:
if self._upgrades_downgrades is None:
result = self._fetch(self.proxy, modules=['upgradeDowngradeHistory'])
if result is None:
self._upgrades_downgrades = pd.DataFrame()
else:
try:
data = result["quoteSummary"]["result"][0]["upgradeDowngradeHistory"]["history"]
if len(data) == 0:
raise YFinanceDataException(f"No upgrade/downgrade history found for {self._symbol}")
df = pd.DataFrame(data)
df.rename(columns={"epochGradeDate": "GradeDate", 'firm': 'Firm', 'toGrade': 'ToGrade', 'fromGrade': 'FromGrade', 'action': 'Action'}, inplace=True)
df.set_index('GradeDate', inplace=True)
df.index = pd.to_datetime(df.index, unit='s')
self._upgrades_downgrades = df
except (KeyError, IndexError):
raise YFinanceDataException(f"Failed to parse json response from Yahoo Finance: {result}")
return self._upgrades_downgrades
@property
def calendar(self) -> dict:
if self._calendar is None:
self._scrape(self.proxy)
self._fetch_calendar()
return self._calendar
def _scrape(self, proxy):
if self._already_scraped:
return
self._already_scraped = True
@staticmethod
def valid_modules():
return quote_summary_valid_modules
# get info and sustainability
json_data = self._data.get_json_data_stores(proxy=proxy)
def _fetch(self, proxy, modules: list):
if not isinstance(modules, list):
raise YFinanceException("Should provide a list of modules, see available modules using `valid_modules`")
modules = ','.join([m for m in modules if m in quote_summary_valid_modules])
if len(modules) == 0:
raise YFinanceException("No valid modules provided, see available modules using `valid_modules`")
params_dict = {"modules": modules, "corsDomain": "finance.yahoo.com", "formatted": "false", "symbol": self._symbol}
try:
quote_summary_store = json_data['QuoteSummaryStore']
except KeyError:
err_msg = "No summary info found, symbol may be delisted"
print('- %s: %s' % (self._data.ticker, err_msg))
result = self._data.get_raw_json(_QUOTE_SUMMARY_URL_ + f"/{self._symbol}", user_agent_headers=self._data.user_agent_headers, params=params_dict, proxy=proxy)
except requests.exceptions.HTTPError as e:
utils.get_yf_logger().error(str(e))
return None
return result
# sustainability
d = {}
try:
if isinstance(quote_summary_store.get('esgScores'), dict):
for item in quote_summary_store['esgScores']:
if not isinstance(quote_summary_store['esgScores'][item], (dict, list)):
d[item] = quote_summary_store['esgScores'][item]
s = pd.DataFrame(index=[0], data=d)[-1:].T
s.columns = ['Value']
s.index.name = '%.f-%.f' % (
s[s.index == 'ratingYear']['Value'].values[0],
s[s.index == 'ratingMonth']['Value'].values[0])
self._sustainability = s[~s.index.isin(
['maxAge', 'ratingYear', 'ratingMonth'])]
except Exception:
pass
self._info = {}
try:
items = ['summaryProfile', 'financialData', 'quoteType',
'defaultKeyStatistics', 'assetProfile', 'summaryDetail']
for item in items:
if isinstance(quote_summary_store.get(item), dict):
self._info.update(quote_summary_store[item])
except Exception:
pass
# For ETFs, provide this valuable data: the top holdings of the ETF
try:
if 'topHoldings' in quote_summary_store:
self._info.update(quote_summary_store['topHoldings'])
except Exception:
pass
try:
if not isinstance(quote_summary_store.get('summaryDetail'), dict):
# For some reason summaryDetail did not give any results. The price dict
# usually has most of the same info
self._info.update(quote_summary_store.get('price', {}))
except Exception:
pass
try:
# self._info['regularMarketPrice'] = self._info['regularMarketOpen']
self._info['regularMarketPrice'] = quote_summary_store.get('price', {}).get(
'regularMarketPrice', self._info.get('regularMarketOpen', None))
except Exception:
pass
try:
self._info['preMarketPrice'] = quote_summary_store.get('price', {}).get(
'preMarketPrice', self._info.get('preMarketPrice', None))
except Exception:
pass
self._info['logo_url'] = ""
try:
if not 'website' in self._info:
self._info['logo_url'] = 'https://logo.clearbit.com/%s.com' % \
self._info['shortName'].split(' ')[0].split(',')[0]
else:
domain = self._info['website'].split(
'://')[1].split('/')[0].replace('www.', '')
self._info['logo_url'] = 'https://logo.clearbit.com/%s' % domain
except Exception:
pass
# Delete redundant info[] keys, because values can be accessed faster
# elsewhere - e.g. price keys. Hope is reduces Yahoo spam effect.
if PRUNE_INFO:
for k in info_retired_keys:
if k in self._info:
del self._info[k]
# InfoDictWrapper will explain how to access above data elsewhere
self._info = InfoDictWrapper(self._info)
# events
try:
cal = pd.DataFrame(quote_summary_store['calendarEvents']['earnings'])
cal['earningsDate'] = pd.to_datetime(
cal['earningsDate'], unit='s')
self._calendar = cal.T
self._calendar.index = utils.camel2title(self._calendar.index)
self._calendar.columns = ['Value']
except Exception as e:
pass
# analyst recommendations
try:
rec = pd.DataFrame(
quote_summary_store['upgradeDowngradeHistory']['history'])
rec['earningsDate'] = pd.to_datetime(
rec['epochGradeDate'], unit='s')
rec.set_index('earningsDate', inplace=True)
rec.index.name = 'Date'
rec.columns = utils.camel2title(rec.columns)
self._recommendations = rec[[
'Firm', 'To Grade', 'From Grade', 'Action']].sort_index()
except Exception:
pass
def _scrape_complementary(self, proxy):
if self._already_scraped_complementary:
def _fetch_info(self, proxy):
if self._already_fetched:
return
self._already_fetched = True
modules = ['financialData', 'quoteType', 'defaultKeyStatistics', 'assetProfile', 'summaryDetail']
result = self._fetch(proxy, modules=modules)
if result is None:
self._info = {}
return
self._already_scraped_complementary = True
self._scrape(proxy)
result["quoteSummary"]["result"][0]["symbol"] = self._symbol
query1_info = next(
(info for info in result.get("quoteSummary", {}).get("result", []) if info["symbol"] == self._symbol),
None,
)
# Most keys that appear in multiple dicts have same value. Except 'maxAge' because
# Yahoo not consistent with days vs seconds. Fix it here:
for k in query1_info:
if "maxAge" in query1_info[k] and query1_info[k]["maxAge"] == 1:
query1_info[k]["maxAge"] = 86400
query1_info = {
k1: v1
for k, v in query1_info.items()
if isinstance(v, dict)
for k1, v1 in v.items()
if v1
}
# recursively format but only because of 'companyOfficers'
def _format(k, v):
if isinstance(v, dict) and "raw" in v and "fmt" in v:
v2 = v["fmt"] if k in {"regularMarketTime", "postMarketTime"} else v["raw"]
elif isinstance(v, list):
v2 = [_format(None, x) for x in v]
elif isinstance(v, dict):
v2 = {k: _format(k, x) for k, x in v.items()}
elif isinstance(v, str):
v2 = v.replace("\xa0", " ")
else:
v2 = v
return v2
for k, v in query1_info.items():
query1_info[k] = _format(k, v)
self._info = query1_info
def _fetch_complementary(self, proxy):
if self._already_fetched_complementary:
return
self._already_fetched_complementary = True
# self._scrape(proxy) # decrypt broken
self._fetch_info(proxy)
if self._info is None:
return
@@ -267,8 +707,7 @@ class Quote:
# pass
#
# For just one/few variable is faster to query directly:
url = "https://query1.finance.yahoo.com/ws/fundamentals-timeseries/v1/finance/timeseries/{}?symbol={}".format(
self._data.ticker, self._data.ticker)
url = f"https://query1.finance.yahoo.com/ws/fundamentals-timeseries/v1/finance/timeseries/{self._symbol}?symbol={self._symbol}"
for k in keys:
url += "&type=" + k
# Request 6 months of data
@@ -280,11 +719,39 @@ class Quote:
json_str = self._data.cache_get(url=url, proxy=proxy).text
json_data = json.loads(json_str)
key_stats = json_data["timeseries"]["result"][0]
if k not in key_stats:
# Yahoo website prints N/A, indicates Yahoo lacks necessary data to calculate
v = None
else:
# Select most recent (last) raw value in list:
v = key_stats[k][-1]["reportedValue"]["raw"]
self._info[k] = v
json_result = json_data.get("timeseries") or json_data.get("finance")
if json_result["error"] is not None:
raise YFinanceException("Failed to parse json response from Yahoo Finance: " + str(json_result["error"]))
for k in keys:
keydict = json_result["result"][0]
if k in keydict:
self._info[k] = keydict[k][-1]["reportedValue"]["raw"]
else:
self.info[k] = None
def _fetch_calendar(self):
# secFilings return too old data, so not requesting it for now
result = self._fetch(self.proxy, modules=['calendarEvents'])
if result is None:
self._calendar = {}
return
try:
self._calendar = dict()
_events = result["quoteSummary"]["result"][0]["calendarEvents"]
if 'dividendDate' in _events:
self._calendar['Dividend Date'] = datetime.datetime.fromtimestamp(_events['dividendDate']).date()
if 'exDividendDate' in _events:
self._calendar['Ex-Dividend Date'] = datetime.datetime.fromtimestamp(_events['exDividendDate']).date()
# splits = _events.get('splitDate') # need to check later, i will add code for this if found data
earnings = _events.get('earnings')
if earnings is not None:
self._calendar['Earnings Date'] = [datetime.datetime.fromtimestamp(d).date() for d in earnings.get('earningsDate', [])]
self._calendar['Earnings High'] = earnings.get('earningsHigh', None)
self._calendar['Earnings Low'] = earnings.get('earningsLow', None)
self._calendar['Earnings Average'] = earnings.get('earningsAverage', None)
self._calendar['Revenue High'] = earnings.get('revenueHigh', None)
self._calendar['Revenue Low'] = earnings.get('revenueLow', None)
self._calendar['Revenue Average'] = earnings.get('revenueAverage', None)
except (KeyError, IndexError):
raise YFinanceDataException(f"Failed to parse json response from Yahoo Finance: {result}")

View File

@@ -0,0 +1,8 @@
daf93e37cbf219cd4c1f3f74ec4551265ec5565b99e8c9322dccd6872941cf13c818cbb88cba6f530e643b4e2329b17ec7161f4502ce6a02bb0dbbe5fc0d0474
ad4d90b3c9f2e1d156ef98eadfa0ff93e4042f6960e54aa2a13f06f528e6b50ba4265a26a1fd5b9cd3db0d268a9c34e1d080592424309429a58bce4adc893c87
e9a8ab8e5620b712ebc2fb4f33d5c8b9c80c0d07e8c371911c785cf674789f1747d76a909510158a7b7419e86857f2d7abbd777813ff64840e4cbc514d12bcae
6ae2523aeafa283dad746556540145bf603f44edbf37ad404d3766a8420bb5eb1d3738f52a227b88283cca9cae44060d5f0bba84b6a495082589f5fe7acbdc9e
3365117c2a368ffa5df7313a4a84988f73926a86358e8eea9497c5ff799ce27d104b68e5f2fbffa6f8f92c1fef41765a7066fa6bcf050810a9c4c7872fd3ebf0
15d8f57919857d5a5358d2082c7ef0f1129cfacd2a6480333dcfb954b7bb67d820abefebfdb0eaa6ef18a1c57f617b67d7e7b0ec040403b889630ae5db5a4dbb
db9630d707a7d0953ac795cd8db1ca9ca6c9d8239197cdfda24b4e0ec9c37eaec4db82dab68b8f606ab7b5b4af3e65dab50606f8cf508269ec927e6ee605fb78
3c895fb5ddcc37d20d3073ed74ee3efad59bcb147c8e80fd279f83701b74b092d503dcd399604c6d8be8f3013429d3c2c76ed5b31b80c9df92d5eab6d3339fce

View File

@@ -22,4 +22,5 @@
_DFS = {}
_PROGRESS_BAR = None
_ERRORS = {}
_TRACEBACKS = {}
_ISINS = {}

View File

@@ -22,36 +22,41 @@
from __future__ import print_function
import datetime as _datetime
import pandas as _pd
from collections import namedtuple as _namedtuple
import pandas as _pd
from .base import TickerBase
from .const import _BASE_URL_
class Ticker(TickerBase):
def __init__(self, ticker, session=None):
super(Ticker, self).__init__(ticker, session=session)
def __init__(self, ticker, session=None, proxy=None):
super(Ticker, self).__init__(ticker, session=session, proxy=proxy)
self._expirations = {}
self._underlying = {}
def __repr__(self):
return 'yfinance.Ticker object <%s>' % self.ticker
return f'yfinance.Ticker object <{self.ticker}>'
def _download_options(self, date=None, proxy=None):
def _download_options(self, date=None):
if date is None:
url = "{}/v7/finance/options/{}".format(
self._base_url, self.ticker)
url = f"{_BASE_URL_}/v7/finance/options/{self.ticker}"
else:
url = "{}/v7/finance/options/{}?date={}".format(
self._base_url, self.ticker, date)
url = f"{_BASE_URL_}/v7/finance/options/{self.ticker}?date={date}"
r = self._data.get(url=url, proxy=proxy).json()
r = self._data.get(url=url, proxy=self.proxy).json()
if len(r.get('optionChain', {}).get('result', [])) > 0:
for exp in r['optionChain']['result'][0]['expirationDates']:
self._expirations[_datetime.datetime.utcfromtimestamp(
exp).strftime('%Y-%m-%d')] = exp
self._underlying = r['optionChain']['result'][0].get('quote', {})
opt = r['optionChain']['result'][0].get('options', [])
return opt[0] if len(opt) > 0 else []
return dict(**opt[0],underlying=self._underlying) if len(opt) > 0 else {}
return {}
def _options2df(self, opt, tz=None):
data = _pd.DataFrame(opt).reindex(columns=[
@@ -76,23 +81,23 @@ class Ticker(TickerBase):
data['lastTradeDate'] = data['lastTradeDate'].dt.tz_convert(tz)
return data
def option_chain(self, date=None, proxy=None, tz=None):
def option_chain(self, date=None, tz=None):
if date is None:
options = self._download_options(proxy=proxy)
options = self._download_options()
else:
if not self._expirations:
self._download_options()
if date not in self._expirations:
raise ValueError(
"Expiration `%s` cannot be found. "
"Available expiration are: [%s]" % (
date, ', '.join(self._expirations)))
f"Expiration `{date}` cannot be found. "
f"Available expirations are: [{', '.join(self._expirations)}]")
date = self._expirations[date]
options = self._download_options(date, proxy=proxy)
options = self._download_options(date)
return _namedtuple('Options', ['calls', 'puts'])(**{
return _namedtuple('Options', ['calls', 'puts', 'underlying'])(**{
"calls": self._options2df(options['calls'], tz=tz),
"puts": self._options2df(options['puts'], tz=tz)
"puts": self._options2df(options['puts'], tz=tz),
"underlying": options['underlying']
})
# ------------------------
@@ -113,12 +118,24 @@ class Ticker(TickerBase):
def mutualfund_holders(self) -> _pd.DataFrame:
return self.get_mutualfund_holders()
@property
def insider_purchases(self) -> _pd.DataFrame:
return self.get_insider_purchases()
@property
def insider_transactions(self) -> _pd.DataFrame:
return self.get_insider_transactions()
@property
def insider_roster_holders(self) -> _pd.DataFrame:
return self.get_insider_roster_holders()
@property
def dividends(self) -> _pd.Series:
return self.get_dividends()
@property
def capital_gains(self):
def capital_gains(self) -> _pd.Series:
return self.get_capital_gains()
@property
@@ -130,25 +147,36 @@ class Ticker(TickerBase):
return self.get_actions()
@property
def shares(self) -> _pd.DataFrame :
def shares(self) -> _pd.DataFrame:
return self.get_shares()
@property
def market_cap(self) -> float:
return self.calc_market_cap()
@property
def info(self) -> dict:
return self.get_info()
@property
def calendar(self) -> _pd.DataFrame:
def fast_info(self):
return self.get_fast_info()
@property
def calendar(self) -> dict:
"""
Returns a dictionary of events, earnings, and dividends for the ticker
"""
return self.get_calendar()
@property
def recommendations(self):
return self.get_recommendations()
@property
def recommendations_summary(self):
return self.get_recommendations_summary()
@property
def upgrades_downgrades(self):
return self.get_upgrades_downgrades()
@property
def earnings(self) -> _pd.DataFrame:
return self.get_earnings()
@@ -213,10 +241,6 @@ class Ticker(TickerBase):
def quarterly_cashflow(self) -> _pd.DataFrame:
return self.quarterly_cash_flow
@property
def recommendations_summary(self):
return self.get_recommendations_summary()
@property
def analyst_price_target(self) -> _pd.DataFrame:
return self.get_analyst_price_target()
@@ -236,9 +260,13 @@ class Ticker(TickerBase):
return tuple(self._expirations.keys())
@property
def news(self):
def news(self) -> list:
return self.get_news()
@property
def trend_details(self) -> _pd.DataFrame:
return self.get_trend_details()
@property
def earnings_trend(self) -> _pd.DataFrame:
return self.get_earnings_trend()

View File

@@ -22,19 +22,21 @@
from __future__ import print_function
from . import Ticker, multi
# from collections import namedtuple as _namedtuple
class Tickers:
def __repr__(self):
return 'yfinance.Tickers object <%s>' % ",".join(self.symbols)
return f"yfinance.Tickers object <{','.join(self.symbols)}>"
def __init__(self, tickers, session=None):
tickers = tickers if isinstance(
tickers, list) else tickers.replace(',', ' ').split()
self.symbols = [ticker.upper() for ticker in tickers]
self.tickers = {ticker:Ticker(ticker, session=session) for ticker in self.symbols}
self.tickers = {ticker: Ticker(ticker, session=session) for ticker in self.symbols}
# self.tickers = _namedtuple(
# "Tickers", ticker_objects.keys(), rename=True
@@ -87,10 +89,4 @@ class Tickers:
return data
def news(self):
collection = {}
for ticker in self.symbols:
collection[ticker] = []
items = Ticker(ticker).news
for item in items:
collection[ticker].append(item)
return collection
return {ticker: [item for item in Ticker(ticker).news] for ticker in self.symbols}

File diff suppressed because it is too large Load Diff

View File

@@ -1 +1 @@
version = "0.2.6"
version = "0.2.38"