mirror of
https://github.com/aykhans/AzSuicideDataVisualization.git
synced 2025-07-04 07:08:05 +00:00
first commit
This commit is contained in:
113
.venv/Lib/site-packages/pyarrow/include/arrow/filesystem/hdfs.h
Normal file
113
.venv/Lib/site-packages/pyarrow/include/arrow/filesystem/hdfs.h
Normal file
@ -0,0 +1,113 @@
|
||||
// Licensed to the Apache Software Foundation (ASF) under one
|
||||
// or more contributor license agreements. See the NOTICE file
|
||||
// distributed with this work for additional information
|
||||
// regarding copyright ownership. The ASF licenses this file
|
||||
// to you under the Apache License, Version 2.0 (the
|
||||
// "License"); you may not use this file except in compliance
|
||||
// with the License. You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing,
|
||||
// software distributed under the License is distributed on an
|
||||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
// KIND, either express or implied. See the License for the
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <memory>
|
||||
#include <string>
|
||||
#include <vector>
|
||||
|
||||
#include "arrow/filesystem/filesystem.h"
|
||||
#include "arrow/io/hdfs.h"
|
||||
#include "arrow/util/uri.h"
|
||||
|
||||
namespace arrow {
|
||||
namespace fs {
|
||||
|
||||
/// Options for the HDFS implementation.
|
||||
struct ARROW_EXPORT HdfsOptions {
|
||||
HdfsOptions() = default;
|
||||
~HdfsOptions() = default;
|
||||
|
||||
/// Hdfs configuration options, contains host, port, driver
|
||||
io::HdfsConnectionConfig connection_config;
|
||||
|
||||
/// Used by Hdfs OpenWritable Interface.
|
||||
int32_t buffer_size = 0;
|
||||
int16_t replication = 3;
|
||||
int64_t default_block_size = 0;
|
||||
|
||||
void ConfigureEndPoint(std::string host, int port);
|
||||
void ConfigureReplication(int16_t replication);
|
||||
void ConfigureUser(std::string user_name);
|
||||
void ConfigureBufferSize(int32_t buffer_size);
|
||||
void ConfigureBlockSize(int64_t default_block_size);
|
||||
void ConfigureKerberosTicketCachePath(std::string path);
|
||||
void ConfigureExtraConf(std::string key, std::string val);
|
||||
|
||||
bool Equals(const HdfsOptions& other) const;
|
||||
|
||||
static Result<HdfsOptions> FromUri(const ::arrow::internal::Uri& uri);
|
||||
static Result<HdfsOptions> FromUri(const std::string& uri);
|
||||
};
|
||||
|
||||
/// HDFS-backed FileSystem implementation.
|
||||
///
|
||||
/// implementation notes:
|
||||
/// - This is a wrapper of arrow/io/hdfs, so we can use FileSystem API to handle hdfs.
|
||||
class ARROW_EXPORT HadoopFileSystem : public FileSystem {
|
||||
public:
|
||||
~HadoopFileSystem() override;
|
||||
|
||||
std::string type_name() const override { return "hdfs"; }
|
||||
HdfsOptions options() const;
|
||||
bool Equals(const FileSystem& other) const override;
|
||||
|
||||
/// \cond FALSE
|
||||
using FileSystem::GetFileInfo;
|
||||
/// \endcond
|
||||
Result<FileInfo> GetFileInfo(const std::string& path) override;
|
||||
Result<std::vector<FileInfo>> GetFileInfo(const FileSelector& select) override;
|
||||
|
||||
Status CreateDir(const std::string& path, bool recursive = true) override;
|
||||
|
||||
Status DeleteDir(const std::string& path) override;
|
||||
|
||||
Status DeleteDirContents(const std::string& path, bool missing_dir_ok = false) override;
|
||||
|
||||
Status DeleteRootDirContents() override;
|
||||
|
||||
Status DeleteFile(const std::string& path) override;
|
||||
|
||||
Status Move(const std::string& src, const std::string& dest) override;
|
||||
|
||||
Status CopyFile(const std::string& src, const std::string& dest) override;
|
||||
|
||||
Result<std::shared_ptr<io::InputStream>> OpenInputStream(
|
||||
const std::string& path) override;
|
||||
Result<std::shared_ptr<io::RandomAccessFile>> OpenInputFile(
|
||||
const std::string& path) override;
|
||||
Result<std::shared_ptr<io::OutputStream>> OpenOutputStream(
|
||||
const std::string& path,
|
||||
const std::shared_ptr<const KeyValueMetadata>& metadata = {}) override;
|
||||
Result<std::shared_ptr<io::OutputStream>> OpenAppendStream(
|
||||
const std::string& path,
|
||||
const std::shared_ptr<const KeyValueMetadata>& metadata = {}) override;
|
||||
|
||||
/// Create a HdfsFileSystem instance from the given options.
|
||||
static Result<std::shared_ptr<HadoopFileSystem>> Make(
|
||||
const HdfsOptions& options, const io::IOContext& = io::default_io_context());
|
||||
|
||||
protected:
|
||||
HadoopFileSystem(const HdfsOptions& options, const io::IOContext&);
|
||||
|
||||
class Impl;
|
||||
std::unique_ptr<Impl> impl_;
|
||||
};
|
||||
|
||||
} // namespace fs
|
||||
} // namespace arrow
|
Reference in New Issue
Block a user