use std::fs::File;
use std::io::Cursor;
use std::num::NonZeroUsize;
use std::path::PathBuf;
pub use arrow::array::StructArray;
use num_traits::pow::Pow;
use polars_core::prelude::*;
use polars_core::utils::accumulate_dataframes_vertical;
use polars_core::POOL;
use rayon::prelude::*;
use crate::mmap::{MmapBytesReader, ReaderBytes};
use crate::ndjson::buffer::*;
use crate::prelude::*;
const NEWLINE: u8 = b'\n';
const RETURN: u8 = b'\r';
const CLOSING_BRACKET: u8 = b'}';
#[must_use]
pub struct JsonLineReader<'a, R>
where
R: MmapBytesReader,
{
reader: R,
rechunk: bool,
n_rows: Option<usize>,
n_threads: Option<usize>,
infer_schema_len: Option<NonZeroUsize>,
chunk_size: NonZeroUsize,
schema: Option<SchemaRef>,
schema_overwrite: Option<&'a Schema>,
path: Option<PathBuf>,
low_memory: bool,
ignore_errors: bool,
}
impl<'a, R> JsonLineReader<'a, R>
where
R: 'a + MmapBytesReader,
{
pub fn with_n_rows(mut self, num_rows: Option<usize>) -> Self {
self.n_rows = num_rows;
self
}
pub fn with_schema(mut self, schema: SchemaRef) -> Self {
self.schema = Some(schema);
self
}
pub fn with_schema_overwrite(mut self, schema: &'a Schema) -> Self {
self.schema_overwrite = Some(schema);
self
}
pub fn with_rechunk(mut self, rechunk: bool) -> Self {
self.rechunk = rechunk;
self
}
pub fn infer_schema_len(mut self, infer_schema_len: Option<NonZeroUsize>) -> Self {
self.infer_schema_len = infer_schema_len;
self
}
pub fn with_n_threads(mut self, n: Option<usize>) -> Self {
self.n_threads = n;
self
}
pub fn with_path<P: Into<PathBuf>>(mut self, path: Option<P>) -> Self {
self.path = path.map(|p| p.into());
self
}
pub fn with_chunk_size(mut self, chunk_size: Option<NonZeroUsize>) -> Self {
if let Some(chunk_size) = chunk_size {
self.chunk_size = chunk_size;
};
self
}
pub fn low_memory(mut self, toggle: bool) -> Self {
self.low_memory = toggle;
self
}
pub fn with_ignore_errors(mut self, ignore_errors: bool) -> Self {
self.ignore_errors = ignore_errors;
self
}
}
impl<'a> JsonLineReader<'a, File> {
pub fn from_path<P: Into<PathBuf>>(path: P) -> PolarsResult<Self> {
let path = resolve_homedir(&path.into());
let f = polars_utils::open_file(&path)?;
Ok(Self::new(f).with_path(Some(path)))
}
}
impl<'a, R> SerReader<R> for JsonLineReader<'a, R>
where
R: MmapBytesReader,
{
fn new(reader: R) -> Self {
JsonLineReader {
reader,
rechunk: true,
n_rows: None,
n_threads: None,
infer_schema_len: Some(NonZeroUsize::new(100).unwrap()),
schema: None,
schema_overwrite: None,
path: None,
chunk_size: NonZeroUsize::new(1 << 18).unwrap(),
low_memory: false,
ignore_errors: false,
}
}
fn finish(mut self) -> PolarsResult<DataFrame> {
let rechunk = self.rechunk;
let reader_bytes = get_reader_bytes(&mut self.reader)?;
let mut json_reader = CoreJsonReader::new(
reader_bytes,
self.n_rows,
self.schema,
self.schema_overwrite,
self.n_threads,
1024, self.chunk_size,
self.low_memory,
self.infer_schema_len,
self.ignore_errors,
)?;
let mut df: DataFrame = json_reader.as_df()?;
if rechunk && df.n_chunks() > 1 {
df.as_single_chunk_par();
}
Ok(df)
}
}
pub(crate) struct CoreJsonReader<'a> {
reader_bytes: Option<ReaderBytes<'a>>,
n_rows: Option<usize>,
schema: SchemaRef,
n_threads: Option<usize>,
sample_size: usize,
chunk_size: NonZeroUsize,
low_memory: bool,
ignore_errors: bool,
}
impl<'a> CoreJsonReader<'a> {
#[allow(clippy::too_many_arguments)]
pub(crate) fn new(
reader_bytes: ReaderBytes<'a>,
n_rows: Option<usize>,
schema: Option<SchemaRef>,
schema_overwrite: Option<&Schema>,
n_threads: Option<usize>,
sample_size: usize,
chunk_size: NonZeroUsize,
low_memory: bool,
infer_schema_len: Option<NonZeroUsize>,
ignore_errors: bool,
) -> PolarsResult<CoreJsonReader<'a>> {
let reader_bytes = reader_bytes;
let mut schema = match schema {
Some(schema) => schema,
None => {
let bytes: &[u8] = &reader_bytes;
let mut cursor = Cursor::new(bytes);
Arc::new(crate::ndjson::infer_schema(&mut cursor, infer_schema_len)?)
},
};
if let Some(overwriting_schema) = schema_overwrite {
let schema = Arc::make_mut(&mut schema);
overwrite_schema(schema, overwriting_schema)?;
}
Ok(CoreJsonReader {
reader_bytes: Some(reader_bytes),
schema,
sample_size,
n_rows,
n_threads,
chunk_size,
low_memory,
ignore_errors,
})
}
fn parse_json(&mut self, mut n_threads: usize, bytes: &[u8]) -> PolarsResult<DataFrame> {
let mut bytes = bytes;
let mut total_rows = 128;
if let Some((mean, std)) = get_line_stats_json(bytes, self.sample_size) {
let line_length_upper_bound = mean + 1.1 * std;
total_rows = (bytes.len() as f32 / (mean - 0.01 * std)) as usize;
if let Some(n_rows) = self.n_rows {
total_rows = std::cmp::min(n_rows, total_rows);
let n_bytes = (line_length_upper_bound * (n_rows as f32)) as usize;
if n_bytes < bytes.len() {
if let Some(pos) = next_line_position_naive_json(&bytes[n_bytes..]) {
bytes = &bytes[..n_bytes + pos]
}
}
}
}
if total_rows <= 128 {
n_threads = 1;
}
let rows_per_thread = total_rows / n_threads;
let max_proxy = bytes.len() / n_threads / 2;
let capacity = if self.low_memory {
usize::from(self.chunk_size)
} else {
std::cmp::min(rows_per_thread, max_proxy)
};
let file_chunks = get_file_chunks_json(bytes, n_threads);
let dfs = POOL.install(|| {
file_chunks
.into_par_iter()
.map(|(start_pos, stop_at_nbytes)| {
let mut buffers = init_buffers(&self.schema, capacity, self.ignore_errors)?;
parse_lines(&bytes[start_pos..stop_at_nbytes], &mut buffers)?;
DataFrame::new(
buffers
.into_values()
.map(|buf| buf.into_series())
.collect::<_>(),
)
})
.collect::<PolarsResult<Vec<_>>>()
})?;
accumulate_dataframes_vertical(dfs)
}
pub fn as_df(&mut self) -> PolarsResult<DataFrame> {
let n_threads = self.n_threads.unwrap_or_else(|| POOL.current_num_threads());
let reader_bytes = self.reader_bytes.take().unwrap();
let mut df = self.parse_json(n_threads, &reader_bytes)?;
if let Some(n_rows) = self.n_rows {
if n_rows < df.height() {
df = df.slice(0, n_rows)
}
}
Ok(df)
}
}
#[inline(always)]
fn parse_impl(
bytes: &[u8],
buffers: &mut PlIndexMap<BufferKey, Buffer>,
scratch: &mut Vec<u8>,
) -> PolarsResult<usize> {
scratch.clear();
scratch.extend_from_slice(bytes);
let n = scratch.len();
let all_good = match n {
0 => true,
1 => scratch[0] == NEWLINE,
2 => scratch[0] == NEWLINE && scratch[1] == RETURN,
_ => {
let value: simd_json::BorrowedValue = simd_json::to_borrowed_value(scratch)
.map_err(|e| polars_err!(ComputeError: "error parsing line: {}", e))?;
match value {
simd_json::BorrowedValue::Object(value) => {
buffers.iter_mut().try_for_each(|(s, inner)| {
match s.0.map_lookup(&value) {
Some(v) => inner.add(v)?,
None => inner.add_null(),
}
PolarsResult::Ok(())
})?;
},
_ => {
buffers.iter_mut().for_each(|(_, inner)| inner.add_null());
},
};
true
},
};
polars_ensure!(all_good, ComputeError: "invalid JSON: unexpected end of file");
Ok(n)
}
fn parse_lines(bytes: &[u8], buffers: &mut PlIndexMap<BufferKey, Buffer>) -> PolarsResult<()> {
let mut buf = vec![];
let iter =
serde_json::Deserializer::from_slice(bytes).into_iter::<Box<serde_json::value::RawValue>>();
for value_result in iter {
match value_result {
Ok(value) => {
let bytes = value.get().as_bytes();
parse_impl(bytes, buffers, &mut buf)?;
},
Err(e) => {
polars_bail!(ComputeError: "error parsing ndjson {}", e)
},
}
}
Ok(())
}
pub(crate) fn next_line_position_naive_json(input: &[u8]) -> Option<usize> {
let pos = memchr::memchr(NEWLINE, input)?;
if pos == 0 {
return Some(1);
}
let is_closing_bracket = input.get(pos - 1) == Some(&CLOSING_BRACKET);
if is_closing_bracket {
Some(pos + 1)
} else {
None
}
}
pub(crate) fn get_line_stats_json(bytes: &[u8], n_lines: usize) -> Option<(f32, f32)> {
let mut lengths = Vec::with_capacity(n_lines);
let mut bytes_trunc;
let n_lines_per_iter = n_lines / 2;
let mut n_read = 0;
let bytes_len = bytes.len();
for offset in [0, (bytes_len as f32 * 0.75) as usize] {
bytes_trunc = &bytes[offset..];
let pos = next_line_position_naive_json(bytes_trunc)?;
if pos >= bytes_len {
return None;
}
bytes_trunc = &bytes_trunc[pos + 1..];
for _ in offset..(offset + n_lines_per_iter) {
let pos = next_line_position_naive_json(bytes_trunc);
if let Some(pos) = pos {
lengths.push(pos);
let next_bytes = &bytes_trunc[pos..];
if next_bytes.is_empty() {
return None;
}
bytes_trunc = next_bytes;
n_read += pos;
} else {
break;
}
}
}
let n_samples = lengths.len();
let mean = (n_read as f32) / (n_samples as f32);
let mut std = 0.0;
for &len in lengths.iter() {
std += (len as f32 - mean).pow(2.0)
}
std = (std / n_samples as f32).sqrt();
Some((mean, std))
}
pub(crate) fn get_file_chunks_json(bytes: &[u8], n_threads: usize) -> Vec<(usize, usize)> {
let mut last_pos = 0;
let total_len = bytes.len();
let chunk_size = total_len / n_threads;
let mut offsets = Vec::with_capacity(n_threads);
for _ in 0..n_threads {
let search_pos = last_pos + chunk_size;
if search_pos >= bytes.len() {
break;
}
let end_pos = match next_line_position_naive_json(&bytes[search_pos..]) {
Some(pos) => search_pos + pos,
None => {
break;
},
};
offsets.push((last_pos, end_pos));
last_pos = end_pos;
}
offsets.push((last_pos, total_len));
offsets
}