Cargo update & some doc tweak

This commit is contained in:
Vincent Prouillet 2018-10-30 15:07:23 +01:00
parent aad12d829f
commit d524482ac1
3 changed files with 225 additions and 203 deletions

406
Cargo.lock generated

File diff suppressed because it is too large Load diff

View file

@ -5,7 +5,7 @@ authors = ["Vincent Prouillet <prouillet.vincent@gmail.com>"]
[dependencies]
tera = "0.11"
base64 = "0.9"
base64 = "0.10"
lazy_static = "1"
pulldown-cmark = "0"
toml = "0.4"

View file

@ -112,7 +112,7 @@ impl Hash for DataSource {
}
fn get_data_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> {
fn get_data_source_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> {
let path_arg = optional_arg!(
String,
args.get("path"),
@ -157,15 +157,15 @@ fn get_output_format_from_args(args: &HashMap<String, Value>, data_source: &Data
}
/// A global function to load data from a data file.
/// Currently the supported formats are json, toml and csv
/// A global function to load data from a file or from a URL
/// Currently the supported formats are json, toml, csv and plain text
pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
let mut headers = header::HeaderMap::new();
headers.insert(header::USER_AGENT, "zola".parse().unwrap());
let client = Arc::new(Mutex::new(Client::builder().build().expect("reqwest client build")));
let result_cache: Arc<Mutex<HashMap<u64, Value>>> = Arc::new(Mutex::new(HashMap::new()));
Box::new(move |args| -> Result<Value> {
let data_source = get_data_from_args(&content_path, &args)?;
let data_source = get_data_source_from_args(&content_path, &args)?;
let file_format = get_output_format_from_args(&args, &data_source)?;
@ -200,23 +200,19 @@ pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
})
}
/// load/parse a json file from the given path and place it into a
/// tera value
/// Parse a JSON string and convert it to a Tera Value
fn load_json(json_data: String) -> Result<Value> {
let json_content: Value = serde_json::from_str(json_data.as_str()).map_err(|e| format!("{:?}", e))?;
return Ok(json_content);
}
/// load/parse a toml file from the given path, and place it into a
/// tera Value
/// Parse a TOML string and convert it to a Tera Value
fn load_toml(toml_data: String) -> Result<Value> {
let toml_content: toml::Value = toml::from_str(&toml_data).map_err(|e| format!("{:?}", e))?;
to_value(toml_content).map_err(|e| e.into())
}
/// Load/parse a csv file from the given path, and place it into a
/// tera Value.
/// Parse a CSV string and convert it to a Tera Value
///
/// An example csv file `example.csv` could be:
/// ```csv
@ -236,7 +232,6 @@ fn load_toml(toml_data: String) -> Result<Value> {
/// ```
fn load_csv(csv_data: String) -> Result<Value> {
let mut reader = Reader::from_reader(csv_data.as_bytes());
let mut csv_map = Map::new();
{
@ -257,7 +252,6 @@ fn load_csv(csv_data: String) -> Result<Value> {
for result in records {
let record = result.unwrap();
let mut elements_array: Vec<Value> = Vec::new();
for e in record.into_iter() {