Cargo update & some doc tweak

This commit is contained in:
Vincent Prouillet 2018-10-30 15:07:23 +01:00
parent aad12d829f
commit d524482ac1
3 changed files with 225 additions and 203 deletions

406
Cargo.lock generated

File diff suppressed because it is too large Load diff

View file

@ -5,7 +5,7 @@ authors = ["Vincent Prouillet <prouillet.vincent@gmail.com>"]
[dependencies] [dependencies]
tera = "0.11" tera = "0.11"
base64 = "0.9" base64 = "0.10"
lazy_static = "1" lazy_static = "1"
pulldown-cmark = "0" pulldown-cmark = "0"
toml = "0.4" toml = "0.4"

View file

@ -112,7 +112,7 @@ impl Hash for DataSource {
} }
fn get_data_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> { fn get_data_source_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> {
let path_arg = optional_arg!( let path_arg = optional_arg!(
String, String,
args.get("path"), args.get("path"),
@ -157,15 +157,15 @@ fn get_output_format_from_args(args: &HashMap<String, Value>, data_source: &Data
} }
/// A global function to load data from a data file. /// A global function to load data from a file or from a URL
/// Currently the supported formats are json, toml and csv /// Currently the supported formats are json, toml, csv and plain text
pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn { pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
let mut headers = header::HeaderMap::new(); let mut headers = header::HeaderMap::new();
headers.insert(header::USER_AGENT, "zola".parse().unwrap()); headers.insert(header::USER_AGENT, "zola".parse().unwrap());
let client = Arc::new(Mutex::new(Client::builder().build().expect("reqwest client build"))); let client = Arc::new(Mutex::new(Client::builder().build().expect("reqwest client build")));
let result_cache: Arc<Mutex<HashMap<u64, Value>>> = Arc::new(Mutex::new(HashMap::new())); let result_cache: Arc<Mutex<HashMap<u64, Value>>> = Arc::new(Mutex::new(HashMap::new()));
Box::new(move |args| -> Result<Value> { Box::new(move |args| -> Result<Value> {
let data_source = get_data_from_args(&content_path, &args)?; let data_source = get_data_source_from_args(&content_path, &args)?;
let file_format = get_output_format_from_args(&args, &data_source)?; let file_format = get_output_format_from_args(&args, &data_source)?;
@ -200,23 +200,19 @@ pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
}) })
} }
/// load/parse a json file from the given path and place it into a /// Parse a JSON string and convert it to a Tera Value
/// tera value
fn load_json(json_data: String) -> Result<Value> { fn load_json(json_data: String) -> Result<Value> {
let json_content: Value = serde_json::from_str(json_data.as_str()).map_err(|e| format!("{:?}", e))?; let json_content: Value = serde_json::from_str(json_data.as_str()).map_err(|e| format!("{:?}", e))?;
return Ok(json_content); return Ok(json_content);
} }
/// load/parse a toml file from the given path, and place it into a /// Parse a TOML string and convert it to a Tera Value
/// tera Value
fn load_toml(toml_data: String) -> Result<Value> { fn load_toml(toml_data: String) -> Result<Value> {
let toml_content: toml::Value = toml::from_str(&toml_data).map_err(|e| format!("{:?}", e))?; let toml_content: toml::Value = toml::from_str(&toml_data).map_err(|e| format!("{:?}", e))?;
to_value(toml_content).map_err(|e| e.into()) to_value(toml_content).map_err(|e| e.into())
} }
/// Load/parse a csv file from the given path, and place it into a /// Parse a CSV string and convert it to a Tera Value
/// tera Value.
/// ///
/// An example csv file `example.csv` could be: /// An example csv file `example.csv` could be:
/// ```csv /// ```csv
@ -236,7 +232,6 @@ fn load_toml(toml_data: String) -> Result<Value> {
/// ``` /// ```
fn load_csv(csv_data: String) -> Result<Value> { fn load_csv(csv_data: String) -> Result<Value> {
let mut reader = Reader::from_reader(csv_data.as_bytes()); let mut reader = Reader::from_reader(csv_data.as_bytes());
let mut csv_map = Map::new(); let mut csv_map = Map::new();
{ {
@ -257,7 +252,6 @@ fn load_csv(csv_data: String) -> Result<Value> {
for result in records { for result in records {
let record = result.unwrap(); let record = result.unwrap();
let mut elements_array: Vec<Value> = Vec::new(); let mut elements_array: Vec<Value> = Vec::new();
for e in record.into_iter() { for e in record.into_iter() {