use std::collections::VecDeque;
use std::path::{Path, PathBuf};
use std::sync::LazyLock;
use arrow::buffer::Buffer;
use polars_core::config;
use polars_core::error::{PolarsResult, polars_bail, to_compute_err};
use polars_utils::pl_str::PlSmallStr;
use polars_utils::plpath::{CloudScheme, PlPath, PlPathRef};
#[cfg(feature = "cloud")]
mod hugging_face;
use crate::cloud::CloudOptions;
#[allow(clippy::bind_instead_of_map)]
pub static POLARS_TEMP_DIR_BASE_PATH: LazyLock<Box<Path>> = LazyLock::new(|| {
(|| {
let verbose = config::verbose();
let path = if let Ok(v) = std::env::var("POLARS_TEMP_DIR").map(PathBuf::from) {
if verbose {
eprintln!("init_temp_dir: sourced from POLARS_TEMP_DIR")
}
v
} else if cfg!(target_family = "unix") {
let id = std::env::var("USER")
.inspect(|_| {
if verbose {
eprintln!("init_temp_dir: sourced $USER")
}
})
.or_else(|_e| {
#[cfg(feature = "file_cache")]
{
std::env::var("HOME")
.inspect(|_| {
if verbose {
eprintln!("init_temp_dir: sourced $HOME")
}
})
.map(|x| blake3::hash(x.as_bytes()).to_hex()[..32].to_string())
}
#[cfg(not(feature = "file_cache"))]
{
Err(_e)
}
});
if let Ok(v) = id {
std::env::temp_dir().join(format!("polars-{v}/"))
} else {
return Err(std::io::Error::other(
"could not load $USER or $HOME environment variables",
));
}
} else if cfg!(target_family = "windows") {
std::env::temp_dir().join("polars/")
} else {
std::env::temp_dir().join("polars/")
}
.into_boxed_path();
if let Err(err) = std::fs::create_dir_all(path.as_ref()) {
if !path.is_dir() {
panic!(
"failed to create temporary directory: {} (path = {:?})",
err,
path.as_ref()
);
}
}
#[cfg(target_family = "unix")]
{
use std::os::unix::fs::PermissionsExt;
let result = (|| {
std::fs::set_permissions(path.as_ref(), std::fs::Permissions::from_mode(0o700))?;
let perms = std::fs::metadata(path.as_ref())?.permissions();
if (perms.mode() % 0o1000) != 0o700 {
std::io::Result::Err(std::io::Error::other(format!(
"permission mismatch: {perms:?}"
)))
} else {
std::io::Result::Ok(())
}
})()
.map_err(|e| {
std::io::Error::new(
e.kind(),
format!(
"error setting temporary directory permissions: {} (path = {:?})",
e,
path.as_ref()
),
)
});
if std::env::var("POLARS_ALLOW_UNSECURED_TEMP_DIR").as_deref() != Ok("1") {
result?;
}
}
std::io::Result::Ok(path)
})()
.map_err(|e| {
std::io::Error::new(
e.kind(),
format!(
"error initializing temporary directory: {e} \
consider explicitly setting POLARS_TEMP_DIR"
),
)
})
.unwrap()
});
pub fn resolve_homedir(path: &dyn AsRef<Path>) -> PathBuf {
let path = path.as_ref();
if path.starts_with("~") {
#[cfg(not(target_family = "wasm"))]
if let Some(homedir) = home::home_dir() {
return homedir.join(path.strip_prefix("~").unwrap());
}
}
path.into()
}
pub fn get_glob_start_idx(path: &[u8]) -> Option<usize> {
memchr::memchr3(b'*', b'?', b'[', path)
}
pub fn expanded_from_single_directory(paths: &[PlPath], expanded_paths: &[PlPath]) -> bool {
paths.len() == 1 && get_glob_start_idx(paths[0].as_ref().strip_scheme().as_bytes()).is_none()
&& {
(
paths[0].as_ref().as_local_path().is_some_and(|p| p.is_dir())
)
|| (
expanded_paths.is_empty() || (paths[0] != expanded_paths[0])
)
}
}
pub fn expand_paths(
paths: &[PlPath],
glob: bool,
hidden_file_prefix: &[PlSmallStr],
#[allow(unused_variables)] cloud_options: &mut Option<CloudOptions>,
) -> PolarsResult<Buffer<PlPath>> {
expand_paths_hive(paths, glob, hidden_file_prefix, cloud_options, false).map(|x| x.0)
}
struct HiveIdxTracker<'a> {
idx: usize,
paths: &'a [PlPath],
check_directory_level: bool,
}
impl HiveIdxTracker<'_> {
fn update(&mut self, i: usize, path_idx: usize) -> PolarsResult<()> {
let check_directory_level = self.check_directory_level;
let paths = self.paths;
if check_directory_level
&& ![usize::MAX, i].contains(&self.idx)
&& (path_idx > 0 && paths[path_idx].as_ref().parent() != paths[path_idx - 1].as_ref().parent())
{
polars_bail!(
InvalidOperation:
"attempted to read from different directory levels with hive partitioning enabled: \
first path: {}, second path: {}",
paths[path_idx - 1].display(),
paths[path_idx].display(),
)
} else {
self.idx = std::cmp::min(self.idx, i);
Ok(())
}
}
}
pub fn expand_paths_hive(
paths: &[PlPath],
glob: bool,
hidden_file_prefix: &[PlSmallStr],
#[allow(unused_variables)] cloud_options: &mut Option<CloudOptions>,
check_directory_level: bool,
) -> PolarsResult<(Buffer<PlPath>, usize)> {
let Some(first_path) = paths.first() else {
return Ok((vec![].into(), 0));
};
let is_cloud = first_path.as_ref().is_cloud_url();
let is_hidden_file = move |path: &PlPath| {
path.as_ref()
.file_name()
.and_then(|x| x.to_str())
.is_some_and(|file_name| {
hidden_file_prefix
.iter()
.any(|x| file_name.starts_with(x.as_str()))
})
};
let mut out_paths = OutPaths {
paths: vec![],
exts: [None, None],
current_idx: 0,
is_hidden_file: &is_hidden_file,
};
let mut hive_idx_tracker = HiveIdxTracker {
idx: usize::MAX,
paths,
check_directory_level,
};
if is_cloud || { cfg!(not(target_family = "windows")) && config::force_async() } {
#[cfg(feature = "cloud")]
{
use polars_utils::_limit_path_len_io_err;
use crate::cloud::object_path_from_str;
if first_path.cloud_scheme() == Some(CloudScheme::Hf) {
let (expand_start_idx, paths) = crate::pl_async::get_runtime().block_in_place_on(
hugging_face::expand_paths_hf(
paths,
check_directory_level,
cloud_options,
glob,
),
)?;
return Ok((paths.into(), expand_start_idx));
}
let format_path = |scheme: &str, bucket: &str, location: &str| {
if is_cloud {
format!("{scheme}://{bucket}/{location}")
} else {
format!("/{location}")
}
};
let expand_path_cloud = |path: PlPathRef<'_>,
cloud_options: Option<&CloudOptions>|
-> PolarsResult<(usize, Vec<PlPath>)> {
crate::pl_async::get_runtime().block_in_place_on(async {
let path_str = path.to_str();
let (cloud_location, store) =
crate::cloud::build_object_store(path, cloud_options, glob).await?;
let prefix = object_path_from_str(&cloud_location.prefix)?;
let out = if !path_str.ends_with("/")
&& (!glob || cloud_location.expansion.is_none())
&& {
path.is_cloud_url() || path.as_local_path().unwrap().is_file()
} {
(
0,
vec![PlPath::from_string(format_path(
cloud_location.scheme,
&cloud_location.bucket,
prefix.as_ref(),
))],
)
} else {
use futures::TryStreamExt;
if let Some(path) = path.as_local_path() {
if !path.is_dir() {
path.metadata()
.map_err(|err| _limit_path_len_io_err(path, err))?;
}
}
let cloud_location = &cloud_location;
let mut paths = store
.try_exec_rebuild_on_err(|store| {
let st = store.clone();
async {
let store = st;
let out = store
.list(Some(&prefix))
.try_filter_map(|x| async move {
let out = (x.size > 0).then(|| {
PlPath::from_string({
format_path(
cloud_location.scheme,
&cloud_location.bucket,
x.location.as_ref(),
)
})
});
Ok(out)
})
.try_collect::<Vec<_>>()
.await?;
Ok(out)
}
})
.await?;
let mut prefix = prefix.to_string();
if path_str.ends_with('/') && !prefix.ends_with('/') {
prefix.push('/')
};
paths.sort_unstable();
(
format_path(
cloud_location.scheme,
&cloud_location.bucket,
prefix.as_ref(),
)
.len(),
paths,
)
};
PolarsResult::Ok(out)
})
};
for (path_idx, path) in paths.iter().enumerate() {
use std::borrow::Cow;
let mut path = Cow::Borrowed(path);
if matches!(
path.cloud_scheme(),
Some(CloudScheme::Http | CloudScheme::Https)
) {
let mut rewrite_aws = false;
#[cfg(feature = "aws")]
if let Some(p) = (|| {
use crate::cloud::CloudConfig;
let p = path.as_ref().as_ref();
let after_scheme = p.strip_scheme();
let bucket_end = after_scheme.find(".s3.")?;
let offset = bucket_end + 4;
let region_end = offset + after_scheme[offset..].find(".amazonaws.com/")?;
if after_scheme[..region_end].contains('/') || after_scheme.contains('?') {
return None;
}
let bucket = &after_scheme[..bucket_end];
let region = &after_scheme[bucket_end + 4..region_end];
let key = &after_scheme[region_end + 15..];
if let CloudConfig::Aws(configs) = cloud_options
.get_or_insert_default()
.config
.get_or_insert_with(|| CloudConfig::Aws(Vec::with_capacity(1)))
{
use object_store::aws::AmazonS3ConfigKey;
if !matches!(configs.last(), Some((AmazonS3ConfigKey::Region, _))) {
configs.push((AmazonS3ConfigKey::Region, region.into()))
}
}
Some(format!("s3://{bucket}/{key}"))
})() {
path = Cow::Owned(PlPath::from_string(p));
rewrite_aws = true;
}
if !rewrite_aws {
out_paths.push(path.into_owned());
hive_idx_tracker.update(0, path_idx)?;
continue;
}
}
let glob_start_idx = get_glob_start_idx(path.to_str().as_bytes());
let path = if glob && glob_start_idx.is_some() {
path.clone()
} else {
let (expand_start_idx, paths) =
expand_path_cloud(path.as_ref().as_ref(), cloud_options.as_ref())?;
out_paths.extend_from_slice(&paths);
hive_idx_tracker.update(expand_start_idx, path_idx)?;
continue;
};
hive_idx_tracker.update(0, path_idx)?;
let iter = crate::pl_async::get_runtime().block_in_place_on(crate::async_glob(
path.as_ref().as_ref(),
cloud_options.as_ref(),
))?;
if is_cloud {
out_paths.extend(iter.into_iter().map(PlPath::from_string));
} else {
out_paths.extend(
iter.iter()
.map(|x| &x[7..])
.map(|s| PlPathRef::new(s).into_owned()),
)
}
}
}
#[cfg(not(feature = "cloud"))]
panic!("Feature `cloud` must be enabled to use globbing patterns with cloud urls.")
} else {
let mut stack = VecDeque::new();
let mut paths_scratch = vec![];
for (path_idx, path) in paths.iter().enumerate() {
let path = path.as_ref();
let path = path.as_local_path().unwrap();
stack.clear();
if path.is_dir() {
let path = path.to_path_buf();
let i = path.to_str().unwrap().len();
hive_idx_tracker.update(i, path_idx)?;
stack.push_back(path.clone());
while let Some(dir) = stack.pop_front() {
let mut last_err = Ok(());
paths_scratch.clear();
paths_scratch.extend(std::fs::read_dir(dir)?.map_while(|x| match x {
Ok(v) => Some(v.path()),
Err(e) => {
last_err = Err(e);
None
},
}));
last_err?;
paths_scratch.sort_unstable();
for path in paths_scratch.drain(..) {
if path.is_dir() {
stack.push_back(path);
} else if path.metadata()?.len() > 0 {
out_paths.push(PlPath::Local(path.into()));
}
}
}
continue;
}
let i = get_glob_start_idx(path.to_str().unwrap().as_bytes());
if glob && i.is_some() {
hive_idx_tracker.update(0, path_idx)?;
let Ok(paths) = glob::glob(path.to_str().unwrap()) else {
polars_bail!(ComputeError: "invalid glob pattern given")
};
for path in paths {
let path = path.map_err(to_compute_err)?;
if !path.is_dir() && path.metadata()?.len() > 0 {
out_paths.push(PlPath::Local(path.into()));
}
}
} else {
hive_idx_tracker.update(0, path_idx)?;
out_paths.push(PlPath::Local(path.into()));
}
}
}
assert_eq!(out_paths.current_idx, out_paths.paths.len());
if expanded_from_single_directory(paths, out_paths.paths.as_slice()) {
if let [Some((_, i1)), Some((_, i2))] = out_paths.exts {
polars_bail!(
InvalidOperation: "directory contained paths with different file extensions: \
first path: {}, second path: {}. Please use a glob pattern to explicitly specify \
which files to read (e.g. 'dir/**/*', 'dir/**/*.parquet')",
&out_paths.paths[i1].display(), &out_paths.paths[i2].display()
)
}
}
return Ok((out_paths.paths.into(), hive_idx_tracker.idx));
struct OutPaths<'a, F: Fn(&PlPath) -> bool> {
paths: Vec<PlPath>,
exts: [Option<(PlSmallStr, usize)>; 2],
current_idx: usize,
is_hidden_file: &'a F,
}
impl<F> OutPaths<'_, F>
where
F: Fn(&PlPath) -> bool,
{
fn push(&mut self, value: PlPath) {
if (self.is_hidden_file)(&value) {
return;
}
let current_idx = &mut self.current_idx;
let exts = &mut self.exts;
Self::update_ext_status(current_idx, exts, value.as_ref());
self.paths.push(value)
}
fn extend(&mut self, values: impl IntoIterator<Item = PlPath>) {
let current_idx = &mut self.current_idx;
let exts = &mut self.exts;
self.paths.extend(
values
.into_iter()
.filter(|x| !(self.is_hidden_file)(x))
.inspect(|x| {
Self::update_ext_status(current_idx, exts, x.as_ref());
}),
)
}
fn extend_from_slice(&mut self, values: &[PlPath]) {
self.extend(values.iter().cloned())
}
fn update_ext_status(
current_idx: &mut usize,
exts: &mut [Option<(PlSmallStr, usize)>; 2],
value: PlPathRef,
) {
let ext = value
.extension()
.map(PlSmallStr::from)
.unwrap_or(PlSmallStr::EMPTY);
if exts[0].is_none() {
exts[0] = Some((ext, *current_idx));
} else if exts[1].is_none() && ext != exts[0].as_ref().unwrap().0 {
exts[1] = Some((ext, *current_idx));
}
*current_idx += 1;
}
}
}
#[cfg(feature = "file_cache")]
pub(crate) fn ensure_directory_init(path: &Path) -> std::io::Result<()> {
let result = std::fs::create_dir_all(path);
if path.is_dir() { Ok(()) } else { result }
}
#[cfg(test)]
mod tests {
use std::path::PathBuf;
use polars_utils::plpath::PlPath;
use super::resolve_homedir;
#[cfg(not(target_os = "windows"))]
#[test]
fn test_resolve_homedir() {
let paths: Vec<PathBuf> = vec![
"~/dir1/dir2/test.csv".into(),
"/abs/path/test.csv".into(),
"rel/path/test.csv".into(),
"/".into(),
"~".into(),
];
let resolved: Vec<PathBuf> = paths.iter().map(|x| resolve_homedir(x)).collect();
assert_eq!(resolved[0].file_name(), paths[0].file_name());
assert!(resolved[0].is_absolute());
assert_eq!(resolved[1], paths[1]);
assert_eq!(resolved[2], paths[2]);
assert_eq!(resolved[3], paths[3]);
assert!(resolved[4].is_absolute());
}
#[cfg(target_os = "windows")]
#[test]
fn test_resolve_homedir_windows() {
let paths: Vec<PathBuf> = vec![
r#"c:\Users\user1\test.csv"#.into(),
r#"~\user1\test.csv"#.into(),
"~".into(),
];
let resolved: Vec<PathBuf> = paths.iter().map(|x| resolve_homedir(x)).collect();
assert_eq!(resolved[0], paths[0]);
assert_eq!(resolved[1].file_name(), paths[1].file_name());
assert!(resolved[1].is_absolute());
assert!(resolved[2].is_absolute());
}
#[test]
fn test_http_path_with_query_parameters_is_not_expanded_as_glob() {
use super::expand_paths;
let path = "https://pola.rs/test.csv?token=bear";
let paths = &[PlPath::new(path)];
let out = expand_paths(paths, true, &[], &mut None).unwrap();
assert_eq!(out.as_ref(), paths);
}
}