Skip to content

NotImplementedError: Parquet writer option(s) ['write.parquet.row-group-size-bytes'] not implemented #1013

@djouallah

Description

@djouallah

Apache Iceberg version

0.7.0 (latest release)

Please describe the bug 🐞

it was working fine, and today, I got this ? using Tabular as a catalog

/usr/local/lib/python3.10/dist-packages/pydantic/main.py:415: UserWarning: Pydantic serializer warnings:
  Expected `TableIdentifier` but got `dict` - serialized value may not be as expected
  return self.__pydantic_serializer__.to_json(
---------------------------------------------------------------------------
NotImplementedError                       Traceback (most recent call last)
<timed exec> in <module>

[/usr/local/lib/python3.10/dist-packages/pyiceberg/table/__init__.py](https://localhost:8080/#) in append(self, df, snapshot_properties)
   1572         """
   1573         with self.transaction() as tx:
-> 1574             tx.append(df=df, snapshot_properties=snapshot_properties)
   1575 
   1576     def overwrite(

3 frames
[/usr/local/lib/python3.10/dist-packages/pyiceberg/io/pyarrow.py](https://localhost:8080/#) in _get_parquet_writer_kwargs(table_properties)
   2288     ]:
   2289         if unsupported_keys := fnmatch.filter(table_properties, key_pattern):
-> 2290             raise NotImplementedError(f"Parquet writer option(s) {unsupported_keys} not implemented")
   2291 
   2292     compression_codec = table_properties.get(TableProperties.PARQUET_COMPRESSION, TableProperties.PARQUET_COMPRESSION_DEFAULT)

NotImplementedError: Parquet writer option(s) ['write.parquet.row-group-size-bytes'] not implemented

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type
    No fields configured for issues without a type.

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions