Pandas version checks
-
[X] I have checked that this issue has not already been reported.
-
[X] I have confirmed this bug exists on the latest version of pandas.
-
[x] I have confirmed this bug exists on the main branch of pandas.
Reproducible Example
import pandas as pd
idx = [0,1,2]
df = pd.DataFrame(0,index=idx, columns=idx)
print(df)
df.to_csv("Test.csv")
df2 = pd.read_csv("Test.csv", header=[0], index_col=[0])
print(df2)
#Test.csv
#All values, including row and column indices, are integers
#,0,1,2
#0,0,0,0
#1,0,0,0
#2,0,0,0
# row and column indices of original df are integers
print(df.columns)
print(df.index)
# column indices of df created from csv are strings
print(df2.columns)
print(df2.index)
# Work-around:
df_temp = pd.read_csv('Test.csv',header=[0], index_col=[0])
df3 = pd.DataFrame(0,index = df_temp.index, columns=[])
for c in df_temp.columns:
newc:int = int(c)
df3[newc] = df_temp[c]
print(df3.columns)
Issue Description
Issue: read_csv() returns the column indices, created as integers, as strings rather than integers. However, the row index is returned as integers.
The programmer assumes the columns are integers because that is how the original df was created, so will get a KeyError when running df[0], when only df['0'] now works.
FYI, the same problem occurs for MultiIndex dataframes: df[('0','0') works but df[(0,0)] generates a KeyError.
This is a difficult bug (unexpected feature?) to find because the difference is not evident when printing the dataframe, only when printing df.columns. The Variable Explorer in Spyder also displays 0 not '0' for the column indices, so I may cross-post this to Spyder.
For database processing, the use of integer column indices is rare, but for those that use the index value as part of the computation, such as X(i+1) = F(X(i)), having integer indices is convenient, if not important. The main thing is that the recreation of row and column indices from csv files are unexpectedly different.
I would assume that the same behaviour could be in other file importers.
A suggestions for removing the confusion: 1. Add a parameter in read_csv to select the dtype of the column (and row) index, just as there are parameters to select the dtype of the data in the columns. 2. Add a parameter to to_csv() to specify the dtype of the indices. This is probably the best solution since it will create csv files that are correctly formated. eg. strings have quotes around them, while integers do not. That way read_csv() has a way to interprete the formating.
Expected Behavior
See Issue Description and run sample code
Installed Versions
INSTALLED VERSIONS
commit : e86ed377639948c64c429059127bcf5b359ab6be python : 3.11.5.final.0 python-bits : 64 OS : Linux OS-release : 6.2.0-37-generic Version : #38~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Thu Nov 2 18:01:13 UTC 2 machine : x86_64 processor : x86_64 byteorder : little LC_ALL : None LANG : en_CA.UTF-8 LOCALE : en_CA.UTF-8
pandas : 2.1.1 numpy : 1.24.3 pytz : 2023.3.post1 dateutil : 2.8.2 setuptools : 68.0.0 pip : 23.2.1 Cython : None pytest : 7.4.0 hypothesis : None sphinx : 5.0.2 blosc : None feather : None xlsxwriter : None lxml.etree : 4.9.3 html5lib : None pymysql : None psycopg2 : None jinja2 : 3.1.2 IPython : 8.15.0 pandas_datareader : None bs4 : 4.12.2 bottleneck : 1.3.5 dataframe-api-compat: None fastparquet : None fsspec : 2023.4.0 gcsfs : None matplotlib : 3.7.2 numba : 0.57.1 numexpr : 2.8.4 odfpy : None openpyxl : 3.0.10 pandas_gbq : None pyarrow : 11.0.0 pyreadstat : None pyxlsb : None s3fs : 2023.4.0 scipy : 1.11.3 sqlalchemy : 1.4.39 tables : 3.8.0 tabulate : 0.8.10 xarray : 2023.6.0 xlrd : None zstandard : 0.19.0 tzdata : 2023.3 qtpy : 2.2.0 pyqt5 : None
Comment From: tuhinsharma121
@thinkerf can I work on it?
Comment From: thinkerf
Tuhin,
Go at it. I haven't had time to look at the code.
Bruce
On 2024-04-01 19:19, Tuhin Sharma wrote:
@thinkerf [1] can I work on it?
-- Reply to this email directly, view it on GitHub [2], or unsubscribe [3]. You are receiving this because you were mentioned.Message ID: @.***>
Links:
[1] https://github.com/thinkerf [2] https://github.com/pandas-dev/pandas/issues/56380#issuecomment-2030956377 [3] https://github.com/notifications/unsubscribe-auth/AJONFMJHK63LQSUWTSTNYB3Y3IITPAVCNFSM6AAAAABALQZFYWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDAMZQHE2TMMZXG4
Comment From: tuhinsharma121
take
Comment From: tuhinsharma121
Definitely its an issue.
Both the following piece of code converts the dataframe to the same CSV format
import pandas as pd
idx = [0, 1, 2]
columns = ["0", "1", "2"]
df = pd.DataFrame(0, index=idx, columns=columns)
print(df)
df.to_csv("Test.csv")
and
import pandas as pd
idx = [0, 1, 2]
columns = [0, 1, 2]
df = pd.DataFrame(0, index=idx, columns=columns)
print(df)
df.to_csv("Test.csv")
give the same output Test.csv
,0,1,2
0,0,0,0
1,0,0,0
2,0,0,0
So for pd.read_csv() it becomes an ambiguous problem.
FWIW I would use the following code snippet as work around
import pandas as pd
import numpy as np
print("Work-around:")
df = pd.read_csv('Test.csv', header=[0], index_col=[0])
df.columns = df.columns.astype(np.int64)
print(df.columns)
Output would be as follows
Work-around:
Index([0, 1, 2], dtype='int64')
I agree with your suggestion. But I shall wait for a pandas team member reviewing this and giving a go ahead.
A suggestions for removing the confusion:
- Add a parameter in read_csv to select the dtype of the column (and row) index, just as there are parameters to select the >dtype of the data in the columns.
- Add a parameter to to_csv() to specify the dtype of the indices. This is probably the best solution since it will create csv >files that are correctly formated. eg. strings have quotes around them, while integers do not. That way read_csv() has a way >to interprete the formating.