With the following code if I use a long query then I get an error. Is there a limit on the query size that I give?
Sorry, but I can't share the exact query that I'm passing. If the same query is given in MySQL workbench then the data is returned correctly. Any advice would be appreciated.
import pandas as pd
connection = pymysql.connect(host = host,
user = user,
db = db,
password = password,
port = port)
df = pd.read_sql(sql_query, connection)
query = '''long query'''
#use pandas to pass SQL query
df = sql_to_df(query)
497 sql, index_col=index_col, params=params,
498 coerce_float=coerce_float, parse_dates=parse_dates,
--> 499 chunksize=chunksize)
1602 if chunksize is not None:
TypeError: 'NoneType' object is not iterable
Using a different tool to query the database (mysql.connector). I was able to find additional insight into the error.
DatabaseError: Execution failed on sql 'SET SESSION group_concat_max_len = 1000000;
After removing that from my query it worked.