oafp reference list of examples

Examples of use of oafp avaiable also in https://ojob.io/oafp-examples.yaml.

πŸ“š Contents

Category Sub-category # Description
AI BedRock 1 List all AWS BedRock models available to use.
AI BedRock 2 Place a prompt to get a list of data using AWS BedRock titan express model.
AI BedRock 3 Place a prompt to get a list of data using an AWS BedRock Llama model.
AI Classification 4 Given a list of news titles with corresponding dates and sources add a category and sub-category field to each
AI Classification 5 Using Google's Gemini model and a list of jar files will try to produce a table categorizing and describing the function of each jar file.
AI Conversation 6 Send multiple requests to an OpenAI model keeping the conversation between interactions and setting the temperature parameter
AI DeepSeek 7 Given a list of files and their corresponding sizes use DeepSeek R1 model to produce a list of 5 files that would be interesting to use to test an ETL tool.
AI Evaluate 8 Given a strace summary of the execution of a command (CMD) use a LLM to describe a reason on the results obtained.
AI Generate data 9 Generates synthetic data making parallel prompts to a LLM model and then cleaning up the result removing duplicate records into a final csv file.
AI Generate data 10 Generates synthetic data using a LLM model and then uses the recorded conversation to generate more data respecting the provided instructions
AI Groq 11 Calculation
AI Mistral 12 List all available LLM models at mistral.ai
AI Ollama 13 Setting up access to Ollama and ask for data to an AI LLM model
AI OpenAI 14 Setting up the OpenAI LLM model and gather the data into a data.json file
AI Prompt 15 Example of generating data from an AI LLM prompt
AI Scaleway 16 Setting up a Scaleway Generative API LLM connection and ask for a list of all Roman emperors.
AI Summarize 17 Use an AI LLM model to summarize the weather information provided in a JSON format
AI TogetherAI 18 List all available LLM models at together.ai with their corresponding type and price components order by the more expensive, for input and output
APIs NASA 19 Markdown table of near Earth objects by name, magnitude, if is potentially hazardous asteroids and corresponding distance
APIs Network 20 Converting the Cloudflare DNS trace info
APIs Network 21 Converting the Google DNS DoH query result
APIs Network 22 Generating a simple map of the current public IP address
APIs Public Holidays 23 Return the public holidays for a given country on a given year
APIs Space 24 How many people are in space and in which craft currently in space
APIs iTunes 25 Search the Apple's iTunes database for a specific term
AWS DynamoDB 26 Given an AWS DynamoDB table 'my-table' will produce a ndjson output with all table items.
AWS DynamoDB 27 Given an AWS DynamoDB table 'my-users' will produce a colored tree output by getting the item for the email key 'scott.tiger@example.com'
AWS EC2 28 Given all AWS EC2 instances in an account produces a table with name, type, vpc and private ip sorted by vpc
AWS EKS 29 Builds an excel spreadsheet with all persistent volumes associated with an AWS EKS 'XYZ' with the corresponding Kubernetes namespace, pvc and pv names
AWS EKS 30 Produce a list of all AWS EKS Kubernetes service accounts, per namespace, which have an AWS IAM role associated with it.
AWS Inspector 31 Given an AWS ECR repository and image tag retrieve the AWS Inspector vulnerabilities and list a quick overview in an Excel spreadsheet
AWS Lambda 32 Prepares a table of AWS Lambda functions with their corresponding main details
AWS Lambda 33 Prepares a table, for a specific AWS Lambda function during a specific time periods, with number of invocations and minimum, average and maximum duration per periods from AWS CloudWatch
AWS RDS Data 34 Given an AWS RDS, postgresql based, database with RDS Data API activated execute the 'analyze' statement for each table on the 'public' schema.
Azure Bing 35 Given an Azure Bing Search API key and a query returns the corresponding search result from Bing.
Channels S3 36 Given a S3 bucket will load data from a previously store data in a provided prefix
Channels S3 37 Given a S3 bucket will save a list of data (the current list of name and versions of OpenAF's oPacks) within a provided prefix
Chart Unix 38 Output a chart with the current Unix load using uptime
DB H2 39 Perform a SQL query over a H2 database.
DB H2 40 Perform queries and DML SQL statements over a H2 databases
DB H2 41 Store the json result of a command into a H2 database table.
DB List 42 List all OpenAF's oPack pre-prepared JDBC drivers
DB Mongo 43 List all records from a specific MongoDB database and collection from a remote Mongo database.
DB SQLite 44 Lists all files in openaf.jar, stores the result in a 'data' table on a SQLite data.db file and performs a query over the stored data.
DB SQLite 45 Perform a query over a database using JDBC.
DB SQLite 46 Store the json result on a SQLite database table.
Diff Envs 47 Given two JSON files with environment variables performs a diff and returns a colored result with the corresponding differences
Diff Lines 48 Performing a diff between two long command lines to spot differences
Diff Path 49 Given two JSON files with the parsed PATH environment variable performs a diff and returns a colored result with the corresponding differences
Docker Containers 50 Output a table with the list of running containers.
Docker Listing 51 List all containers with the docker-compose project, service name, file, id, name, image, creation time, status, networks and ports.
Docker Listing 52 List all containers with their corresponding labels parsed and sorted.
Docker Network 53 Output a table with the docker networks info.
Docker Registry 54 List all a table of docker container images repository and corresponding tags of a private registry.
Docker Registry 55 List all the docker container image repositories of a private registry.
Docker Registry 56 List all the docker container image repository tags of a private registry.
Docker Stats 57 Output a table with the docker stats broken down for each value.
Docker Storage 58 Output a table with the docker volumes info.
Docker Volumes 59 Given a list of docker volumes will remove all of them, if not in use, in parallel for faster execution.
ElasticSearch Cluster 60 Get an ElasticSearch/OpenSearch cluster nodes overview
ElasticSearch Cluster 61 Get an ElasticSearch/OpenSearch cluster per host data allocation
ElasticSearch Cluster 62 Get an ElasticSearch/OpenSearch cluster settings flat
ElasticSearch Cluster 63 Get an ElasticSearch/OpenSearch cluster settings non-flatted
ElasticSearch Cluster 64 Get an ElasticSearch/OpenSearch cluster stats per node
ElasticSearch Cluster 65 Get an overview of an ElasticSearch/OpenSearch cluster health
ElasticSearch Indices 66 Get an ElasticSearch/OpenSearch count per index
ElasticSearch Indices 67 Get an ElasticSearch/OpenSearch indices overview
ElasticSearch Indices 68 Get an ElasticSearch/OpenSearch settings for a specific index
GIT History 69 Give a GIT repository will retrieve the current log history and parse it to an Excel (XLS) file.
GPU Nvidia 70 Builds a grid with two charts providing a visualization over a Nvidia GPU usage and the corresponding memory usage for a specific GPU_IDX (gpu index)
GPU Nvidia 71 Get current Nvidia per-gpu usage
Generic Arrays 72 Converting an array of strings into an array of maps
Generic Avro 73 Given an Avro data file outputs it's corresponding statistics
Generic Avro 74 Given an Avro data file outputs the correspoding schema
Generic Avro 75 Reads an Avro data file as input
Generic Avro 76 Write an Avro data file as an output
Generic BOM 77 Given a container image use syft to generate a table with a corresponding bill-of-materials (BOM) table with each identified artifact name, version, type, language, found path and maven group & id (if applicable).
Generic Base64 78 Encode/decode data (or text-like files) to/from gzip base64 representation for easier packing and transport.
Generic CSV 79 Given a BOM CSV with multiple fields (7 in total) return a new CSV with just 3 of the original fields.
Generic Commands 80 Given an input array with phone numbers will run parallel output commands, calling ojob.io/telco/phoneNumber, for each entry effectively building an output from those multiple command executions.
Generic End of life 81 List the versions of a given product with the corresponding end-of-life dates (using endofline.date API)
Generic Excel 82 Building an Excel file with the AWS IPv4 and IPv6 ranges (1).
Generic Excel 83 Building an Excel file with the AWS IPv4 and IPv6 ranges (2).
Generic Excel 84 Building an Excel file with the AWS IPv4 and IPv6 ranges (3).
Generic Excel 85 Processes each json file in /some/data creating and updating the data.xlsx file with a sheet for each file.
Generic Excel 86 Store and retrieve data from an Excel spreadsheet
Generic HTML 87 Generate a HTML with table of emoticons/emojis by category, group, name, unicode and html code.
Generic HTML 88 Given an input file, in a specific language (e.g. yaml, json, bash, etc…), output an HTML representation with syntax highlighting.
Generic Hex 89 Outputs an hexadecimal representation of the characters of the file provided allowing to adjust how many per line/row.
Generic JWT 90 Generates an output JWT (JSON Web Token) given the provided input claims signed with a provided secret.
Generic JWT 91 Given an input JWT (JSON Web Token) converts it to a human readable format.
Generic List files 92 After listing all files and folders recursively producing a count table by file extension.
Generic RSS 93 Builds an HTML file with the current linked news titles, publication date and source from Google News RSS.
Generic RSS 94 Example of generating a HTML list of titles, links and publication dates from a RSS feed
Generic RSS 95 Generates a HTML page with the current news from Google News, order by date, and opens a browser with it.
Generic RSS 96 Parses the Slashdot's RSS feed news into a quick clickable HTML page in a browser
Generic Reverse 97 Given a text file revert the line ordering
Generic Set 98 Given two json files, with arrays of component versions, generate a table with the difference on one of the sides.
Generic Set 99 Given two json files, with arrays of component versions, generate a table with the union of the two sides.
Generic Template 100 Given a meal name will search 'The Meal DB' site for the corresponding recipe and render a markdown HTML of the corresponding recipe.
Generic Text 101 Get a json with lyrics of a song.
Generic Text 102 Search a word in the English dictionary returning phonetic, meanings, synonyms, antonyms, etc.
Generic URL 103 Given an URL to a resource on a website determine how long ago is was modified given the data provided by the server.
Generic YAML 104 Given an YAML file with a data array composed of maps with fields 'c', 's', 'd' and 'e' filter by any record where any field doesn't have contents.
GitHub GIST 105 Using GitHub's GIST functionality retrieves and parses an oAFp examples YAML file with the template and the corresponding data.
GitHub Releases 106 Builds a table of GitHub project releases
GitHub Releases 107 Parses the latest GitHub project release markdown notes
Grid Java 108 Parses a Java hsperf data + the current rss java process memory into a looping grid.
Grid Java 109 Parses a Java hsperf data into a looping grid.
Grid Kubernetes 110 Displays a continuous updating grid with a line chart with the number of CPU throtlles and bursts recorded in the Linux cgroup cpu stats of a container running in Kubernetes and the source cpu.stats data
Grid Mac 111 Shows a grid with the Mac network metrics and 4 charts for in, out packets and in, out bytes
Grid Mac 112 Shows a grid with the Mac storage metrics and 4 charts for read, write IOPS and read, write bytes per second
Grid Unix 113 On an Unix/Linux system supporting 'ps' output formats %cpu and %mem, will output a chart with the percentage of cpu and memory usage of a provided pid (e.g. 12345)
JSON Schemas Lists 114 Get a list of JSON schemas from Schema Store catalog
Java Certificates 115 Given a Java keystore will obtain a list of certificates and output them order by the ones that will expire first.
Kubernetes Base64 116 Given a Kubernetes Config Map or Secret with binary data, retrieves it and stores it locally in a binary file.
Kubernetes Containers 117 Parse the Linux cgroup cpu stats on a container running in Kubernetes
Kubernetes Helm 118 Given an Helm release name and the corresponding namespace will produce a table with the timestamps when the corresponding Helm chart hooks have started and completed for the lastest execution and the corresponding phase.
Kubernetes Kubectl 119 Build a table of the images 'cached' in all Kubernetes nodes using Kubectl and, additionally, provide a summary of the total size per node.
Kubernetes Kubectl 120 Build an output table with Kubernetes pods with namespace, pod name, container name and corresponding resources using kubectl
Kubernetes Kubectl 121 Build an output table with Kubernetes pods with node, namespace, pod name, container name and corresponding resources using kubectl
Kubernetes Kubectl 122 Executes a recursive file list find command in a specific pod, namespace and path converting the result into a table.
Kubernetes Kubectl 123 Given the list of all Kubernetes objects will produce a list of objects per namespace, kind, apiVersiom, creation timestamp, name and owner.
Kubernetes Kubectl 124 List of Kubernetes CPU, memory and storage stats per node using kubectl
Kubernetes Kubectl 125 List of Kubernetes pods per namespace and kind using kubectl
Kubernetes Kubectl 126 Produces a list of pods' containers per namespace with the corresponding images and assigned nodes.
Kubernetes Kubectl 127 Using kubectl with the appropriate permissions check the filesystem available, capacity and used bytes and inodes on each node of the Kubernetes cluster.
Kubernetes PVC 128 Produces a table with all Kubernetes persistent volume claims (PVCs) in use by pods.
Mac Activity 129 Uses the Mac terminal command 'last' output to build an activity table with user, tty, from, login-time and logout-time
Mac Brew 130 List all the packages and corresponding versions installed in a Mac by brew.
Mac Chart 131 On a Mac OS produce a looping chart with the total percentage of current CPU usage.
Mac Info 132 Get a list of the current logged users in Mac OS
Mac Info 133 Parses the current Mac OS hardware information
Mac Info 134 Parses the current Mac OS overview information
Mac Safari 135 Get a list of all Mac OS Safari bookmarks into a CSV file.
Mac Tunnelblink 136 In a Mac OS with Tunnelblink, if you want to copy all your OpenVPN configurations into ovpn files.
Markdown Tables 137 For an input markdown file, parse all tables, transform it to JSON and output as a colored table
Network ASN 138 Retrieve an IP to ASN list list and converts it to ndjson
Network ASN 139 Retrieve the list of ASN number and names from RIPE and transforms it to a CSV.
Network Latency 140 Given a host and a port will display a continuously updating line chart with network latency, in ms, between the current device and the target host and port
Ollama List models 141 Parses the list of models currently in an Ollama deployment
OpenAF Channels 142 Copy the json result of a command into an etcd database using OpenAF's channels
OpenAF Channels 143 Getting all data stored in an etcd database using OpenAF's channels
OpenAF Channels 144 Given a Prometheus database will query for a specific metric (go_memstats_alloc_bytes), during a defined period, every 5 seconds (step) will produce a static chart with the corresponding metric values.
OpenAF Channels 145 Perform a query to a metric & label, with a start and end time, to a Prometheus server using OpenAF's channels
OpenAF Channels 146 Retrieve all keys stores in a H2 MVStore file using OpenAF's channels
OpenAF Channels 147 Store and retrieve data from a Redis database
OpenAF Channels 148 Store and retrieve data from a RocksDB database
OpenAF Channels 149 Store the json results of a command into a H2 MVStore file using OpenAF's channels
OpenAF Flags 150 List the current values of OpenAF/oAFp internal flags
OpenAF Network 151 Gets all the DNS host addresses for a provided domain and ensures that the output is always a list
OpenAF Network 152 List all MX (mail servers) network addresses from the current DNS server for a hostname using OpenAF
OpenAF Network 153 List all network addresses returned from the current DNS server for a hostname using OpenAF
OpenAF OS 154 Current OS information visible to OpenAF
OpenAF OS 155 Using OpenAF parse the current environment variables
OpenAF OpenVPN 156 Using OpenAF code to perform a more complex parsing of the OpenVPN status data running on an OpenVPN container (nmaguiar/openvpn) called 'openvpn'
OpenAF SFTP 157 Generates a file list with filepath, size, permissions, create and last modified time from a SFTP connection with user and password
OpenAF SFTP 158 Generates a file list with filepath, size, permissions, create and last modified time from a SFTP connection with user, private key and password
OpenAF TLS 159 List the TLS certificates of a target host with a sorted alternative names using OpenAF
OpenAF oJob.io 160 Parses ojob.io/news results into a clickable news title HMTL page.
OpenAF oJob.io 161 Retrieves the list of oJob.io's jobs and filters which start by 'ojob.io/news' to display them in a rectangle
OpenAF oPacks 162 Given a folder of expanded oPacks folders will process each folder .package.yaml file and join each corresponding oPack name and dependencies into a sinlge output map.
OpenAF oPacks 163 Listing all currently accessible OpenAF's oPacks
OpenAF oafp 164 Filter the OpenAF's oafp examples list by a specific word in the description
OpenAF oafp 165 List the OpenAF's oafp examples by category, sub-category and description
OpenAF oafp 166 Produce a colored table with all the current oafp input and output formats supported.
OpenVPN List 167 When using the container nmaguiar/openvpn it's possible to convert the list of all clients order by expiration/end date
QR Encode JSON 168 Given a JSON input encode and decote it from a QR-code png file.
QR Read QR-code 169 Given a QR-code png file output the corresponding contents.
QR URL 170 Generate a QR-code for a provided URL.
Unix Activity 171 Uses the Linux command 'last' output to build a table with user, tty, from and period of activity for Debian based Linuxs
Unix Activity 172 Uses the Linux command 'last' output to build a table with user, tty, from and period of activity for RedHat based Linuxs
Unix Alpine 173 List all installed packages in an Alpine system
Unix Ask 174 Unix bash script to ask for a path and choose between filetypes to perform an unix find command.
Unix Compute 175 Parses the Linux /proc/cpuinfo into an array
Unix Debian/Ubuntu 176 List all installed packages in a Debian/Ubuntu system
Unix Envs 177 Converts the Linux envs command result into a table of environment variables and corresponding values
Unix Files 178 Converting the Linux's /etc/os-release to SQL insert statements.
Unix Files 179 Converting the Unix's syslog into a json output.
Unix Files 180 Executes a recursive file list find command converting the result into a table.
Unix Files 181 Parses the Linux /etc/passwd to a table order by uid and gid.
Unix Generic 182 Creates, in unix, a data.ndjson file where each record is formatted from json files in /some/data
Unix Memory map 183 Given an Unix process will output a table with process's components memory address, size in bytes, permissions and owner
Unix Network 184 Loop over the current Linux active network connections
Unix Network 185 Parse the Linux 'arp' command output
Unix Network 186 Parse the Linux 'ip tcp_metrics' command
Unix Network 187 Parse the result of the Linux route command
Unix OpenSuse 188 List all installed packages in an OpenSuse system or zypper based system
Unix RedHat 189 List all installed packages in a RedHat system or rpm based system (use rpm –querytags to list all fields available)
Unix Storage 190 Converting the Unix's df output
Unix Storage 191 Parses the result of the Unix ls command
Unix SystemCtl 192 Converting the Unix's systemctl list-timers
Unix SystemCtl 193 Converting the Unix's systemctl list-units
Unix SystemCtl 194 Converting the Unix's systemctl list-units into an overview table
Unix Threads 195 Given an unix process id (pid) loop a table with its top 25 most cpu active threads
Unix UBI 196 List all installed packages in an UBI system
Unix named 197 Converts a Linux's named log, for client queries, into a CSV
Unix strace 198 Given a strace unix command will produce a summary table of the system calls invoked including a small line chart of the percentage of time of each.
VSCode Extensions 199 Check a Visual Studio Code (vscode) extension (vsix) manifest.
Windows Network 200 Output a table with the current route table using Windows' PowerShell
Windows Network 201 Output a table with the list of network interfaces using Windows' PowerShell
Windows PnP 202 Output a table with USB/PnP devices using Windows' PowerShell
Windows Storage 203 Output a table with the attached disk information using Windows' PowerShell
XML Maven 204 Given a Maven pom.xml parses the XML content to a colored table ordering by the fields groupId and artifactId.
nAttrMon Plugs 205 Given a nAttrMon config folder, with YAML files, produce a summary table with the each plug (yaml file) execFrom definition.

πŸ“— Examples


1

πŸ“– AI | BedRock

List all AWS BedRock models available to use.

export OAFP_MODEL="(type: bedrock, options: ())"
oafp in=llmmodels out=ctable path="[].{id: modelId, arn: modelArn}" libs="@AWS/aws.js" data="()"

2

πŸ“– AI | BedRock

Place a prompt to get a list of data using AWS BedRock titan express model.

export OAFP_MODEL="(type: bedrock, timeout: 900000, options: (model: 'amazon.titan-text-express-v1', temperature: 0, params: (textGenerationConfig: (maxTokenCount: 1024))))"
oafp in=llm libs="@AWS/aws.js" data="list of primary color names" getlist=true out=ctable

3

πŸ“– AI | BedRock

Place a prompt to get a list of data using an AWS BedRock Llama model.

export OAFP_MODEL="(type: bedrock, timeout: 900000, options: (model: 'eu.meta.llama3-2-3b-instruct-v1:0', region: eu-central-1, temperature: 0.1, params: ('top_p': 0.9, 'max_gen_len': 512)))"
oafp in=llm libs="@AWS/aws.js" data="produce a list of european countries with the country name and capital" getlist=true out=ctable

4

πŸ“– AI | Classification

Given a list of news titles with corresponding dates and sources add a category and sub-category field to each

export OAFP_MODEL="(type: ollama, model: 'llama3', url: 'http://ollama.local', timeout: 900000, temperature: 0)"
# get 10 news titles
RSS="https://news.google.com/rss" && oafp url="$RSS" path="rss.channel.item[].{title:title,date:pubDate,source:source._}" from="sort(-date)" out=json sql="select * limit 5" > news.json
# add category and sub-category
oafp news.json llmcontext="list a news titles with date and source" llmprompt="keeping the provided title, date and source add a category and sub-category fields to the provided list" out=json > newsCategorized.json
oafp newsCategorized.json getlist=true out=ctable

5

πŸ“– AI | Classification

Using Google's Gemini model and a list of jar files will try to produce a table categorizing and describing the function of each jar file.

# export OAFP_MODEL=(type: gemini, model: gemini-1.5-flash, key: '...', timeout: 900000, temperature: 0)
oafp in=ls data=lib path="[?ends_with(filename,'.jar')].filename" llmcontext="jar list" llmprompt="add a 'category', 'subcategory' and 'description' to each listed jar file" getlist=true out=json | oafp sql="select * order by category, subcategory" out=ctable

6

πŸ“– AI | Conversation

Send multiple requests to an OpenAI model keeping the conversation between interactions and setting the temperature parameter

export OAFP_MODEL="(type: openai, model: 'gpt-3.5-turbo-0125', key: ..., timeout: 900000, temperature: 0)"
echo "List all countries in the european union" | oafp in=llm out=ctree llmconversation=cvst.json
echo "Add the corresponding country capital" | oafp in=llm out=ctree llmconversation=cvst.json
rm cvst.json

7

πŸ“– AI | DeepSeek

Given a list of files and their corresponding sizes use DeepSeek R1 model to produce a list of 5 files that would be interesting to use to test an ETL tool.

# export OAFP_MODEL="(type: ollama, model: 'deepseek-r1:8b', url: 'https://ollama.local', timeout: 10000000, temperature: 0)"
oafp in=ls data="." path="[].{path:canonicalPath,size:size}" llmcontext="list of local files with path and size" llmprompt="output a json array with the suggestion of 5 data files that would be interesting to use to test an ETL tool" getlist=true out=ctable

8

πŸ“– AI | Evaluate

Given a strace summary of the execution of a command (CMD) use a LLM to describe a reason on the results obtained.

export OAFP_MODEL="(type: openai, model: 'gpt-4o-mini', timeout: 900000, temperature: 0)"
CMD="strace --tips" && strace -c -o /tmp/result.txt $CMD ; cat /tmp/result.txt ; oafp /tmp/result.txt in=raw llmcontext="strace execution json summary" llmprompt="given the provided strace execution summary produce a table-based analysis with a small description of each system call and a bullet point conclusion summary" out=md ; rm /tmp/result.txt

9

πŸ“– AI | Generate data

Generates synthetic data making parallel prompts to a LLM model and then cleaning up the result removing duplicate records into a final csv file.

# Set the LLM model to use
export OAFP_MODEL="(type: openai, url: 'https://api.scaleway.ai', key: '111-222-333', model: 'llama-3.1-70b-instruct', headers: (Content-Type: application/json))"
# Run 5 parallel prompts to generate data
oafp data="()" path="range(\`5\`)" out=cmd outcmdtmpl=true outcmd="oafp data='generate a list of 10 maps each with firstName, lastName, city and country' in=llm out=ndjson" > data.ndjson
# Clean-up generate data removing duplicates
oafp data.ndjson in=ndjson ndjsonjoin=true removedups=true out=csv > data.csv

10

πŸ“– AI | Generate data

Generates synthetic data using a LLM model and then uses the recorded conversation to generate more data respecting the provided instructions

export OAFP_MODEL="(type: ollama, model: 'llama3', url: 'https://models.local', timeout: 900000)"
oafp in=llm llmconversation=conversation.json data="Generate #5 synthetic transaction record data with the following attributes: transaction id - a unique alphanumeric code; date - a date in YYYY-MM-DD format; amount - a dollar amount between 10 and 1000; description - a brief description of the transaction" getlist=true out=ndjson > data.ndjson
oafp in=llm llmconversation=conversation.json data="Generate 5 more records" getlist=true out=ndjson >> data.ndjson
oafp data.ndjson ndjsonjoin=true out=ctable sql="select * order by transaction_id"

11

πŸ“– AI | Groq

Calculation

export OAFP_MODEL="(type: openai, model: 'llama3-70b-8192', key: '...', url: 'https://api.groq.com/openai', timeout: 900000, temperature: 0)"
oafp in=llm data="how much does light take to travel from Tokyo to Osaka in ms; return a 'time_in_ms' and a 'reasoning'"

12

πŸ“– AI | Mistral

List all available LLM models at mistral.ai

# export OAFP_MODEL="(type: openai, model: 'llama3', url: 'https://api.mistral.ai', key: '...', timeout: 900000, temperature: 0)"
oafp in=llmmodels data="()" path="sort([].id)"

13

πŸ“– AI | Ollama

Setting up access to Ollama and ask for data to an AI LLM model

export OAFP_MODEL="(type: ollama, model: 'mistral', url: 'https://models.local', timeout: 900000)"
echo "Output a JSON array with 15 cities where each entry has the 'city' name, the estimated population and the corresponding 'country'" | oafp input=llm output=json > data.json
oafp data.json output=ctable sql="select * order by population desc"

14

πŸ“– AI | OpenAI

Setting up the OpenAI LLM model and gather the data into a data.json file

export OAFP_MODEL="(type: openai, model: gpt-3.5-turbo, key: ..., timeout: 900000)"
echo "list all United Nations secretaries with their corresponding 'name', their mandate 'begin date', their mandate 'end date' and their corresponding secretary 'numeral'" | oafp input=llm output=json > data.json

15

πŸ“– AI | Prompt

Example of generating data from an AI LLM prompt

export OAFP_MODEL="(type: openai, model: 'gpt-3.5-turbo', key: '...', timeout: 900000)"
oafp in=llm data="produce a list of 25 species of 'flowers' with their english and latin name and the continent where it can be found" out=json > data.json

16

πŸ“– AI | Scaleway

Setting up a Scaleway Generative API LLM connection and ask for a list of all Roman emperors.

OAFP_MODEL="(type: openai, url: 'https://api.scaleway.ai', key: '111-222-333', model: 'llama-3.1-70b-instruct', headers: (Content-Type: application/json))"
oafp data="list all roman emperors" in=llm getlist=true out=ctable

17

πŸ“– AI | Summarize

Use an AI LLM model to summarize the weather information provided in a JSON format

export OAFP_MODEL="(type: openai, model: 'gpt-3.5-turbo', key: '...', timeout: 900000)"
oafp url="https://wttr.in?format=j2" llmcontext="current and forecast weather" llmprompt="produce a summary of the current and forecasted weather" out=md

18

πŸ“– AI | TogetherAI

List all available LLM models at together.ai with their corresponding type and price components order by the more expensive, for input and output

# export OAFP_MODEL="(type: openai, model: 'meta-llama/Meta-Llama-3-70B', url: 'https://api.together.xyz', key: '...', timeout: 9000000, temperature: 0)"
oafp in=llmmodels data="()" path="[].{id:id,name:display_name,type:type,ctxLen:context_length,priceHour:pricing.hourly,priceIn:pricing.input,priceOut:pricing.output,priceBase:pricing.base,priceFineTune:pricing.finetune}" sql="select * order by priceIn desc,priceOut desc" out=ctable

19

πŸ“– APIs | NASA

Markdown table of near Earth objects by name, magnitude, if is potentially hazardous asteroids and corresponding distance

curl -s "https://api.nasa.gov/neo/rest/v1/feed?API_KEY=DEMO_KEY" | oafp path="near_earth_objects" maptoarray=true output=json | oafp path="[0][].{name:name,magnitude:absolute_magnitude_h,hazardous:is_potentially_hazardous_asteroid,distance:close_approach_data[0].miss_distance.kilometers}" sql="select * order by distance" output=mdtable

20

πŸ“– APIs | Network

Converting the Cloudflare DNS trace info

curl -s https://1.1.1.1/cdn-cgi/trace | oafp in=ini out=ctree

21

πŸ“– APIs | Network

Converting the Google DNS DoH query result

DOMAIN=yahoo.com && oafp path=Answer from="sort(data)" out=ctable url="https://8.8.8.8/resolve?name=$DOMAIN&type=a"

22

πŸ“– APIs | Network

Generating a simple map of the current public IP address

curl -s https://ifconfig.co/json | oafp flatmap=true out=map

23

πŸ“– APIs | Public Holidays

Return the public holidays for a given country on a given year

COUNTRY=US && YEAR=2024 && oafp url="https://date.nager.at/api/v2/publicholidays/$YEAR/$COUNTRY" path="[].{date:date,localName:localName,name:name}" out=ctable

24

πŸ“– APIs | Space

How many people are in space and in which craft currently in space

curl -s http://api.open-notify.org/astros.json | oafp path="people" sql="select \"craft\", count(1) \"people\" group by \"craft\"" output=ctable

25

πŸ“– APIs | iTunes

Search the Apple's iTunes database for a specific term

TRM="Mozart" && oafp url="https://itunes.apple.com/search?term=$TRM" out=ctree

26

πŸ“– AWS | DynamoDB

Given an AWS DynamoDB table 'my-table' will produce a ndjson output with all table items.

# opack install AWS
oafp libs="@AWS/aws.js" in=ch inch="(type: dynamo, options: (region: us-west-1, tableName: my-table))" inchall=true data="__"  out=ndjson

27

πŸ“– AWS | DynamoDB

Given an AWS DynamoDB table 'my-users' will produce a colored tree output by getting the item for the email key 'scott.tiger@example.com'

# opack install AWS
oafp in=ch inch="(type: dynamo, options: (region: eu-west-1, tableName: my-users))" data="(email: scott-tiger@example.com)" libs="@AWS/aws.js" out=ctree

28

πŸ“– AWS | EC2

Given all AWS EC2 instances in an account produces a table with name, type, vpc and private ip sorted by vpc

aws ec2 describe-instances | oafp path="Reservations[].Instances[].{name:join('',Tags[?Key=='Name'].Value),type:InstanceType,vpc:VpcId,ip:PrivateIpAddress} | sort_by(@, &vpc)" output=ctable

29

πŸ“– AWS | EKS

Builds an excel spreadsheet with all persistent volumes associated with an AWS EKS 'XYZ' with the corresponding Kubernetes namespace, pvc and pv names

# sudo yum install -y fontconfig
aws ec2 describe-volumes | oafp path="Volumes[?Tags[?Key=='kubernetes.io/cluster/XYZ']|[0].Value=='owned'].{VolumeId:VolumeId,Name:Tags[?Key=='Name']|[0].Value,KubeNS:Tags[?Key=='kubernetes.io/created-for/pvc/namespace']|[0].Value,KubePVC:Tags[?Key=='kubernetes.io/created-for/pvc/name']|[0].Value,KubePV:Tags[?Key=='kubernetes.io/created-for/pv/name']|[0].Value,AZ:AvailabilityZone,Size:Size,Type:VolumeType,CreateTime:CreateTime,State:State,AttachTime:join(',',nvl(Attachments[].AttachTime,from_slon('[]'))[]),InstanceId:join(',',nvl(Attachments[].InstanceId,from_slon('[]'))[])}" from="sort(KubeNS,KubePVC)" out=xls xlsfile=xyz_pvs.xlsx

30

πŸ“– AWS | EKS

Produce a list of all AWS EKS Kubernetes service accounts, per namespace, which have an AWS IAM role associated with it.

oafp cmd="kubectl get sa -A -o json" path="items[].{ serviceAccount: metadata.name, ns: metadata.namespace, iamRole: nvl(metadata.annotations.\"eks.amazonaws.com/role-arn\", 'n/a') }" sql="select * where iamRole <> 'n/a'" out=ctable

31

πŸ“– AWS | Inspector

Given an AWS ECR repository and image tag retrieve the AWS Inspector vulnerabilities and list a quick overview in an Excel spreadsheet

REPO=my/image && TAG=1.2.3 && aws inspector2 list-findings --filter "{\"ecrImageRepositoryName\":[{\"comparison\":\"EQUALS\",\"value\":\"$REPO\"}],\"ecrImageTags\":[{\"comparison\":\"EQUALS\",\"value\":\"$TAG\"}]}" --output json | oafp path="findings[].{title:title,severity:severity,lastObservedAt:lastObservedAt,firstObservedAt:firstObservedAt,lastObservedAt:lastObservedAt,fixAvailable:fixAvailable,where:join(', ',(packageVulnerabilityDetails.vulnerablePackages[].nvl(filePath, name)))}" out=xls

32

πŸ“– AWS | Lambda

Prepares a table of AWS Lambda functions with their corresponding main details

aws lambda list-functions | oafp path="Functions[].{Name:FunctionName,Runtime:Runtime,Arch:join(',',Architectures),Role:Role,MemorySize:MemorySize,EphStore:EphemeralStorage.Size,CodeSize:CodeSize,LastModified:LastModified}" from="sort(Name)" out=ctable

33

πŸ“– AWS | Lambda

Prepares a table, for a specific AWS Lambda function during a specific time periods, with number of invocations and minimum, average and maximum duration per periods from AWS CloudWatch

export _SH="aws cloudwatch get-metric-statistics --namespace AWS/Lambda --start-time 2024-03-01T00:00:00Z --end-time 2024-04-01T00:00:00Z --period 3600 --dimensions Name=FunctionName,Value=my-function"
$_SH --statistics Sum --metric-name Invocations  | oafp path="Datapoints[].{ts:Timestamp,cnt:Sum}" out=ndjson > data.ndjson
$_SH --statistics Average --metric-name Duration | oafp path="Datapoints[].{ts:Timestamp,avg:Average}" out=ndjson >> data.ndjson
$_SH --statistics Minimum --metric-name Duration | oafp path="Datapoints[].{ts:Timestamp,min:Minimum}" out=ndjson >> data.ndjson
$_SH --statistics Maximum --metric-name Duration | oafp path="Datapoints[].{ts:Timestamp,max:Maximum}" out=ndjson >> data.ndjson
oafp data.ndjson ndjsonjoin=true opath="[].{ts:ts,cnt:nvl(cnt,\`0\`),min:nvl(min,\`0\`),avg:nvl(avg,\`0\`),max:nvl(max,\`0\`)}" out=json | oafp isql="select \"ts\",max(\"cnt\") \"cnt\",max(\"min\") \"min\",max(\"avg\") \"avg\",max(\"max\") \"max\" group by \"ts\" order by \"ts\"" opath="[].{ts:ts,cnt:cnt,min:from_ms(min,'(abrev:true,pad:true)'),avg:from_ms(avg,'(abrev:true,pad:true)'),max:from_ms(max,'(abrev:true,pad:true)')}" out=ctable

34

πŸ“– AWS | RDS Data

Given an AWS RDS, postgresql based, database with RDS Data API activated execute the 'analyze' statement for each table on the 'public' schema.

SECRET="arn:aws:secretsmanager:eu-west-1:123456789000:secret:dbname-abc123" \
DB="arn:aws:rds:eu-west-1:123456789000:cluster:dbname" \
REGION="eu-west-1" \
oafp libs=AWS in=awsrdsdata awssecret=$SECRET awsdb=$DB awsregion=$REGION \
data="select table_schema, table_name, table_type from information_schema.tables where table_schema='public' and table_type='BASE TABLE'" \
path="formattedRecords[].{libs:'AWS',in:'awsrdsdata',awssecret:'$SECRET',awsdb:'$DB',awsregion:'$REGION',data:concat(concat('analyze \"', table_name),'\"')}" \
out=json | oafp in=oafp

35

πŸ“– Azure | Bing

Given an Azure Bing Search API key and a query returns the corresponding search result from Bing.

QUERY="OpenAF" && KEY="12345" && curl -X GET "https://api.bing.microsoft.com/v7.0/search?q=$QUERY" -H "Ocp-Apim-Subscription-Key: $KEY" | oafp out=ctree

36

πŸ“– Channels | S3

Given a S3 bucket will load data from a previously store data in a provided prefix

# opack install s3
oafp libs="@S3/s3.js" in=ch inch="(type: s3, options: (s3url: 'https://play.min.io:9000', s3accessKey: abc123, s3secretKey: 'xyz123', s3bucket: test, s3prefix: data, multifile: true, gzip: true))" data="()" inchall=true out=ctable

37

πŸ“– Channels | S3

Given a S3 bucket will save a list of data (the current list of name and versions of OpenAF's oPacks) within a provided prefix

# opack install s3
oafp libs="@S3/s3.js" -v path="openaf.opacks" out=ch ch="(type: s3, options: (s3url: 'https://play.min.io:9000', s3accessKey: abc123, s3secretKey: 'xyz123', s3bucket: test, s3prefix: data, multifile: true, gzip: true))" chkey=name

38

πŸ“– Chart | Unix

Output a chart with the current Unix load using uptime

oafp cmd="uptime" in=raw path="replace(trim(@), '.+ ([\d\.]+),? ([\d\.]+),? ([\d\.]+)\$', '', '\$1|\$2|\$3').split(@,'|')" out=chart chart="dec2 [0]:green:load -min:0" loop=1 loopcls=true

39

πŸ“– DB | H2

Perform a SQL query over a H2 database.

echo "select * from \"data\"" | oafp in=db indbjdbc="jdbc:h2:./data" indbuser=sa indbpass=sa out=ctable

40

πŸ“– DB | H2

Perform queries and DML SQL statements over a H2 databases

# Dump data into a table
oafp data="[(id: 1, val: x)|(id: 2, val: y)|(id: 3, val: z)]" in=slon out=db dbjdbc="jdbc:h2:./data" dbuser=sa dbpass=sa dbtable=data
# Copy data to another new table
ESQL='create table newdata as select * from "data"' && oafp in=db indbjdbc="jdbc:h2:./data" indbuser=sa indbpass=sa indbexec=true data="$ESQL"
# Drop the original table
ESQL='drop table "data"' && oafp in=db indbjdbc="jdbc:h2:./data" indbuser=sa indbpass=sa indbexec=true data="$ESQL"
# Output data from the new table
SQL='select * from newdata' && oafp in=db indbjdbc="jdbc:h2:./data" indbuser=sa indbpass=sa data="$SQL" out=ctable

41

πŸ“– DB | H2

Store the json result of a command into a H2 database table.

oaf -c "\$o(listFilesRecursive('.'),{__format:'json'})" | oafp out=db dbjdbc="jdbc:h2:./data" dbuser=sa dbpass=sa dbtable=data

42

πŸ“– DB | List

List all OpenAF's oPack pre-prepared JDBC drivers

oafp in=oaf data="getOPackRemoteDB()" maptoarray=true opath="[].{name:name,description:description,version:version}" from="sort(name)" out=ctable

43

πŸ“– DB | Mongo

List all records from a specific MongoDB database and collection from a remote Mongo database.

# opack install mongo
oafp libs="@Mongo/mongo.js" in=ch inch="(type: mongo, options: (database: default, collection: collection, url: 'mongodb://a.server:27017'))" inchall=true path="[].delete(@,'_id')" data="()"

44

πŸ“– DB | SQLite

Lists all files in openaf.jar, stores the result in a 'data' table on a SQLite data.db file and performs a query over the stored data.

# ojob ojob.io/db/getDriver op=install db=sqlite
opack install jdbc-sqlite
# Dump data into a 'data' table
oafp in=ls data=openaf.jar out=db dbjdbc="jdbc:sqlite:data.db" dblib=sqlite dbtable="data"
# Gets stats over the dump data from the 'data' table
SQL='select count(1) numberOfFiles, round(avg("size")) avgSize, sum("size") totalSize from "data"' && oafp in=db indbjdbc="jdbc:sqlite:data.db" indblib=sqlite data="$SQL" out=ctable

45

πŸ“– DB | SQLite

Perform a query over a database using JDBC.

# ojob ojob.io/db/getDriver op=install db=sqlite
opack install jdbc-sqlite
echo "select * from data" | oafp in=db indbjdbc="jdbc:sqlite:data.db" indbtable=data indblib=sqlite out=ctable

46

πŸ“– DB | SQLite

Store the json result on a SQLite database table.

# ojob ojob.io/db/getDriver op=install db=sqlite
opack install jdbc-sqlite
oaf -c "\$o(listFilesRecursive('.'),{__format:'json'})" | oafp out=db dbjdbc="jdbc:sqlite:data.db" dbtable=data dblib=sqlite

47

πŸ“– Diff | Envs

Given two JSON files with environment variables performs a diff and returns a colored result with the corresponding differences

env | oafp in=ini out=json > data1.json
# change one or more environment variables
env | oafp in=ini out=json > data2.json
oafp in=oafp data="[(file: data1.json)|(file: data2.json)]" diff="(a:'[0]',b:'[1]')" color=true

48

πŸ“– Diff | Lines

Performing a diff between two long command lines to spot differences

oafp diff="(a:before,b:after)" diffchars=true in=yaml color=true
# as stdin enter
before: URL="http://localhost:9090" && METRIC="go_memstats_alloc_bytes" && TYPE="bytes" && LABELS="job=\"prometheus\"" && START="2024-06-18T20:00:00Z" && END="2024-06-18T20:15:00Z" && STEP=5 && echo "{query:'max($METRIC{$LABELS})',start:'$START',end:'$END',step:$STEP}" | oafp in=ch inch="(type:prometheus,options:(urlQuery:'$URL'))" inchall=true out=json | oafp path="[].set(@, 'main').map(&{metric:'$METRIC',job:get('main').metric.job,timestamp:to_date(mul([0],\`1000\`)),value:to_number([1])}, values) | []" out=schart schart="$TYPE '[].value':green:$METRIC -min:0"

after: URL="http://localhost:9090" && METRIC="go_memstats_alloc_bytes" && TYPE="bytes" && LABELS="job=\"prometheus\"" && START="2024-06-18T20:00:00Z" && END="2024-06-18T20:15:00Z" && STEP=5 && echo "{query:'max($METRIC{$LABELS})',start:'$START',end:'$END',step:$STEP}" | oafp in=ch inch="(type:prometheus,options:(urlQuery:'$URL'))" inchall=true out=json | oafp path="[].set(@, 'main').map(&{metric:'$METRIC',job:get('main').metric.job,timestamp:to_date([0]),value:to_number([1])}, values) | []" out=schart schart="$TYPE '[].value':green:$METRIC -min:0"
#
# Ctrl^D 
# as the result the difference between before and after will appear as red characters

49

πŸ“– Diff | Path

Given two JSON files with the parsed PATH environment variable performs a diff and returns a colored result with the corresponding differences

env | oafp in=ini path="PATH.split(@,':')" out=json > data1.json
# export PATH=$PATH:/some/new/path
env | oafp in=ini path="PATH.split(@,':')" out=json > data2.json
oafp in=oafp data="[(file: data1.json)|(file: data2.json)]" diff="(a:'sort([0])',b:'sort([1])')" color=true

50

πŸ“– Docker | Containers

Output a table with the list of running containers.

oafp cmd="docker ps --format json" input=ndjson ndjsonjoin=true path="[].{id:ID,name:Names,state:State,image:Image,networks:Networks,ports:Ports,Status:Status}" sql="select * order by networks,state,name" output=ctable

51

πŸ“– Docker | Listing

List all containers with the docker-compose project, service name, file, id, name, image, creation time, status, networks and ports.

docker ps -a --format=json | oafp in=ndjson ndjsonjoin=true out=ctree path="[].insert(@,'Labels',sort_by(split(Labels,',')[].split(@,'=').{key:[0],value:[1]},&key))" out=json | oafp path="[].{dcProject:nvl(Labels[?key=='com.docker.compose.project']|[0].value,''),dcService:nvl(Labels[?key=='com.docker.compose.service']|[0].value,''),ID:ID,Names:Names,Image:Image,Created:RunningFor,Status:Status,Ports:Ports,Networks:Networks,dcFile:nvl(Labels[?key=='com.docker.compose.project.config_files']|[0].value,'')}" sql="select * order by dcProject,dcService,Networks,Names" out=ctable

52

πŸ“– Docker | Listing

List all containers with their corresponding labels parsed and sorted.

docker ps -a --format=json | oafp in=ndjson ndjsonjoin=true out=ctree path="[].insert(@,'Labels',sort_by(split(Labels,',')[].split(@,'=').{key:[0],value:[1]},&key))" out=ctree

53

πŸ“– Docker | Network

Output a table with the docker networks info.

docker network ls --format json | oafp in=ndjson ndjsonjoin=true out=ctable

54

πŸ“– Docker | Registry

List all a table of docker container images repository and corresponding tags of a private registry.

# opack install DockerRegistry
# check more options with 'oafp libs=dockerregistry help=dockerregistry' 
oafp libs=dockerregistry in=registryrepos data="()" inregistryurl=http://localhost:5000 inregistrytags=true out=ctable

55

πŸ“– Docker | Registry

List all the docker container image repositories of a private registry.

# opack install DockerRegistry
# check more options with 'oafp libs=dockerregistry help=dockerregistry' 
oafp libs=dockerregistry data="()" in=registryrepos inregistryurl=http://localhost:5000

56

πŸ“– Docker | Registry

List all the docker container image repository tags of a private registry.

# opack install DockerRegistry
# check more options with 'oafp libs=dockerregistry help=dockerregistry' 
oafp libs=dockerregistry data="library/nginx" in=registrytags inregistryurl=http://localhost:5000

57

πŸ“– Docker | Stats

Output a table with the docker stats broken down for each value.

oafp cmd="docker stats --no-stream" in=lines linesvisual=true linesjoin=true out=ctree path="[].{containerId:\"CONTAINER ID\",pids:PIDS,name:\"NAME\",cpuPerc:\"CPU %\",memory:\"MEM USAGE / LIMIT\",memPerc:\"MEM %\",netIO:\"NET I/O\",blockIO:\"BLOCK I/O\"}|[].{containerId:containerId,pids:pids,name:name,cpuPerc:replace(cpuPerc,'%','',''),memUsage:from_bytesAbbr(split(memory,' / ')[0]),memLimit:from_bytesAbbr(split(memory,' / ')[1]),memPerc:replace(memPerc,'%','',''),netIn:from_bytesAbbr(split(netIO,' / ')[0]),netOut:from_bytesAbbr(split(netIO,' / ')[1]),blockIn:from_bytesAbbr(split(blockIO,' / ')[0]),blockOut:from_bytesAbbr(split(blockIO,' / ')[1])}" out=ctable

58

πŸ“– Docker | Storage

Output a table with the docker volumes info.

docker volume ls --format json | oafp in=ndjson ndjsonjoin=true out=ctable

59

πŸ“– Docker | Volumes

Given a list of docker volumes will remove all of them, if not in use, in parallel for faster execution.

docker volume ls --format json | oafp in=ndjson ndjsonjoin=true path="[].Name" out=cmd outcmd="docker volume rm {}" outcmdparam=true

60

πŸ“– ElasticSearch | Cluster

Get an ElasticSearch/OpenSearch cluster nodes overview

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cat/nodes?format=json" $ES_EXTRA | oafp sql="select * order by ip" out=ctable

61

πŸ“– ElasticSearch | Cluster

Get an ElasticSearch/OpenSearch cluster per host data allocation

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cat/allocation?format=json&bytes=b" $ES_EXTRA | oafp sql="select * order by host" out=ctable

62

πŸ“– ElasticSearch | Cluster

Get an ElasticSearch/OpenSearch cluster settings flat

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cluster/settings?include_defaults=true&flat_settings=true" $ES_EXTRA | oafp out=ctree

63

πŸ“– ElasticSearch | Cluster

Get an ElasticSearch/OpenSearch cluster settings non-flatted

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cluster/settings?include_defaults=true" $ES_EXTRA | oafp out=ctree

64

πŸ“– ElasticSearch | Cluster

Get an ElasticSearch/OpenSearch cluster stats per node

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_nodes/stats/indices/search" $ES_EXTRA | oafp out=ctree

65

πŸ“– ElasticSearch | Cluster

Get an overview of an ElasticSearch/OpenSearch cluster health

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cat/health?format=json" $ES_EXTRA | oafp out=ctable

66

πŸ“– ElasticSearch | Indices

Get an ElasticSearch/OpenSearch count per index

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/kibana_sample_data_flights/_count" $ES_EXTRA | oafp

67

πŸ“– ElasticSearch | Indices

Get an ElasticSearch/OpenSearch indices overview

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/_cat/indices?format=json&bytes=b" $ES_EXTRA | oafp sql="select * order by index" out=ctable

68

πŸ“– ElasticSearch | Indices

Get an ElasticSearch/OpenSearch settings for a specific index

export ES_URL=http://elastic.search:9200
export ES_EXTRA="--insecure"
curl -s "$ES_URL/kibana_sample_data_flights/_settings" $ES_EXTRA | oafp out=ctree

69

πŸ“– GIT | History

Give a GIT repository will retrieve the current log history and parse it to an Excel (XLS) file.

opack install plugin-XLS
git log --pretty=format:'{c:"%H",a:"%an",d:"%ad",m:"%s"}' | oafp in=ndjson ndjsonjoin=true path="[].{commit:c,author:a,date:to_date(d),message:m}" out=xls outfile=data.xlsx xlsopen=false

70

πŸ“– GPU | Nvidia

Builds a grid with two charts providing a visualization over a Nvidia GPU usage and the corresponding memory usage for a specific GPU_IDX (gpu index)

GPU_IDX=0 &&oafp cmd="nvidia-smi --query-gpu=index,name,memory.total,memory.used,memory.free,utilization.gpu --format=csv,nounits | oafp in=lines path=\"map(&trim(@),split(@,',')).join(',',@)\"" in=csv path="[$GPU_IDX].{memTotal:mul(to_number(\"memory.total [MiB]\"),\`1024\`),memUsed:mul(to_number(\"memory.used [MiB]\"),\`1024\`),memFree:mul(to_number(\"memory.free [MiB]\"),\`1024\`),gpuUse:to_number(\"utilization.gpu [%]\")}" out=grid grid="[[(title: usage, type: chart, obj: 'int gpuUse:green:usage -min:0 -max:100')]|[(title: memory, type: chart, obj: 'bytes memTotal:red:total memUsed:cyan:used -min:0')]]" loopcls=true loop=1

71

πŸ“– GPU | Nvidia

Get current Nvidia per-gpu usage

nvidia-smi --query-gpu=index,name,memory.total,memory.used,memory.free,utilization.gpu --format=csv,nounits | oafp in=lines path="map(&trim(@),split(@,',')).join(',',@)" | oafp in=csv out=ctable

72

πŸ“– Generic | Arrays

Converting an array of strings into an array of maps

oafp -v path="java.params[].insert(from_json('{}'), 'param', @).insert(@, 'len', length(param))"

73

πŸ“– Generic | Avro

Given an Avro data file outputs it's corresponding statistics

# opack install avro
oafp libs=avro data.avro inavrostats=true

74

πŸ“– Generic | Avro

Given an Avro data file outputs the correspoding schema

# opack install avro
oafp libs=avro data.avro inavroschema=true

75

πŸ“– Generic | Avro

Reads an Avro data file as input

# opack install avro
oafp data.avro libs=avro out=ctable

76

πŸ“– Generic | Avro

Write an Avro data file as an output

# opack install avro
oafp data.json libs=avro out=avro avrofile=data.avro

77

πŸ“– Generic | BOM

Given a container image use syft to generate a table with a corresponding bill-of-materials (BOM) table with each identified artifact name, version, type, language, found path and maven group & id (if applicable).

IMAGE=openaf/oaf:t8 && oafp cmd="syft scan registry:$IMAGE -o syft-json" path="artifacts[].{name:name,version:version,type:type,language:language,paths:join(',',locations[].path),groupId:nvl(metadata.pomProperties.groupId,'n/a'),artifactId:nvl(metadata.pomProperties.artifactId,'n/a')}" sql="select * order by type, name, version" out=ctable

78

πŸ“– Generic | Base64

Encode/decode data (or text-like files) to/from gzip base64 representation for easier packing and transport.

# encode a data file to a gzip base64 representation
oafp data.json out=gb64json > data.gb64
# decode a gzip base64 representation back into a data file
oafp data.gb64 in=gb64json out=json > data.json

79

πŸ“– Generic | CSV

Given a BOM CSV with multiple fields (7 in total) return a new CSV with just 3 of the original fields.

# Check current fields (filter for first 5 records)
oafp bom.csv sql="select * limit 5" out=ctable
# Test with just 3 fields (filter for first 5 records)
oafp bom.csv sql="select name, version, group limit 5" out=ctable
# Finally produce the new CSV with just 3 fields
oafp bom.csv sql="select name, version, group" out=csv > new_bom.csv

80

πŸ“– Generic | Commands

Given an input array with phone numbers will run parallel output commands, calling ojob.io/telco/phoneNumber, for each entry effectively building an output from those multiple command executions.

oafp data="[(p:911234567)|(p:+18004564567)]" in=slon out=cmd outcmdtmpl=true outcmd="ojob ojob.io/telco/phoneNumber country=PT number= -json" | oafp in=ndjson ndjsonjoin=true path="[].phoneInfo" out=ctree

81

πŸ“– Generic | End of life

List the versions of a given product with the corresponding end-of-life dates (using endofline.date API)

PRODUCT=jetty && oafp url="https://endoflife.date/api/$PRODUCT.json" out=ctable sql="select * order by releaseDate"

82

πŸ“– Generic | Excel

Building an Excel file with the AWS IPv4 and IPv6 ranges (1).

curl https://ip-ranges.amazonaws.com/ip-ranges.json > ip-ranges.json

83

πŸ“– Generic | Excel

Building an Excel file with the AWS IPv4 and IPv6 ranges (2).

oafp ip-ranges.json path=prefixes out=xls xlsfile=aws-ip-ranges.xlsx xlssheet=ipv4

84

πŸ“– Generic | Excel

Building an Excel file with the AWS IPv4 and IPv6 ranges (3).

oafp ip-ranges.json path=ipv6_prefixes out=xls xlsfile=aws-ip-ranges.xlsx xlssheet=ipv6

85

πŸ“– Generic | Excel

Processes each json file in /some/data creating and updating the data.xlsx file with a sheet for each file.

find /some/data -name "*.json" | xargs -I '{}' /bin/sh -c 'oafp file={} output=xls xlsfile=data.xlsx xlsopen=false xlssheet=$(echo {} | sed "s/.*\/\(.*\)\.json/\1/g" )'

86

πŸ“– Generic | Excel

Store and retrieve data from an Excel spreadsheet

# Storing data
oafp cmd="oaf -c \"sprint(listFilesRecursive('/usr/bin'))\"" out=xls xlsfile=data.xlsx
# Retrieve data
oafp in=xls file=data.xlsx xlscol=A xlsrow=1 out=pjson

87

πŸ“– Generic | HTML

Generate a HTML with table of emoticons/emojis by category, group, name, unicode and html code.

oafp url="https://emojihub.yurace.pro/api/all" path="[].{category:category,group:group,name:name,len:length(unicode),unicode:join('<br>',unicode),htmlCode:replace(join('<br>', htmlCode),'&','g','&amp;'),emoji:join('', htmlCode)}" out=mdtable | oafp in=md out=html

88

πŸ“– Generic | HTML

Given an input file, in a specific language (e.g. yaml, json, bash, etc…), output an HTML representation with syntax highlighting.

OUT=yaml && FILE=data.yaml && oafp file=$FILE in=raw outkey=data out=json | oafp out=template templatetmpl=true template="\`\`\`$OUT\n}\n\`\`\`" | oafp in=md out=html

89

πŸ“– Generic | Hex

Outputs an hexadecimal representation of the characters of the file provided allowing to adjust how many per line/row.

oafp some.file in=rawhex inrawhexline=15 out=ctable

90

πŸ“– Generic | JWT

Generates an output JWT (JSON Web Token) given the provided input claims signed with a provided secret.

oafp data="(claims: (c1: a1, c2: a2))" out=jwt jwtsecret=this_is_my_own_very_long_signature
# you can check it adding "| oafp in=jwt"
# you can also verify the signature adding instead "| oafp in=jwt injwtsecret=this_is_my_own_very_long_signature injwtverify=true" which will add a __verified boolean field.

91

πŸ“– Generic | JWT

Given an input JWT (JSON Web Token) converts it to a human readable format.

echo "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzdWIiOiIxMjM0NTY3ODkwIiwibmFtZSI6IkpvaG4gRG9lIiwiaWF0IjoxNTE2MjM5MDIyfQ.SflKxwRJSMeKKF2QT4fwpMeJf36POk6yJV_adQssw5c" | oafp in=jwt

92

πŸ“– Generic | List files

After listing all files and folders recursively producing a count table by file extension.

FPATH="git/ojob.io" && oafp in=ls lsrecursive=true data="$FPATH" path="[].insert(@,'extension',if(isFile && index_of(filename,'.')>'0',replace(filename,'^.+\.([^\.]+)$','','\$1'),'<dir>'))" from="countBy(extension)" out=json | oafp from="sort(-_count)" out=ctable

93

πŸ“– Generic | RSS

Builds an HTML file with the current linked news titles, publication date and source from Google News RSS.

RSS="https://news.google.com/rss" && oafp url="$RSS" path="rss.channel.item[].{title:replace(t(@,'[]()'),'\|','g','\\|'),date:pubDate,source:source._}" from="sort(-date)" out=mdtable | oafp in=md out=html

94

πŸ“– Generic | RSS

Example of generating a HTML list of titles, links and publication dates from a RSS feed

oafp url="https://blog.google/rss" path="rss.channel.item" sql="select title, link, pubDate" output=html

95

πŸ“– Generic | RSS

Generates a HTML page with the current news from Google News, order by date, and opens a browser with it.

oafp url="https://news.google.com/rss" path="rss.channel.item[].{title:title,link:link,date:pubDate,source:source._}" out=template templatetmpl=true template="<html><body><h1>Current Main News</h1><ul><li><a href='' target='_blank'></a> <br><small> | Source: </small></li></ul></body></html>" sql="select * order by date desc" | oafp in=raw out=html

96

πŸ“– Generic | RSS

Parses the Slashdot's RSS feed news into a quick clickable HTML page in a browser

RSS="http://rss.slashdot.org/Slashdot/slashdot" && oafp url="$RSS" path="RDF.item[].{title:replace(t(@,'[]()'),'\|','g','\\|'),department:department,date:date}" from="sort(-date)" out=mdtable | oafp in=md out=html

97

πŸ“– Generic | Reverse

Given a text file revert the line ordering

cat somefile.txt | oafp in=lines linesjoin=true path="reverse([]).join('\n',@)"

98

πŸ“– Generic | Set

Given two json files, with arrays of component versions, generate a table with the difference on one of the sides.

oafp data="[(file: versions-latest.json)|(file: versions-build.json)]" in=oafp set="(a:'[0]',b:'[1]')" setop=diffb out=ctable

99

πŸ“– Generic | Set

Given two json files, with arrays of component versions, generate a table with the union of the two sides.

oafp data="[(file: versions-latest.json)|(file: versions-build.json)]" in=oafp set="(a:'[0]',b:'[1]')" setop=union out=ctable

100

πŸ“– Generic | Template

Given a meal name will search 'The Meal DB' site for the corresponding recipe and render a markdown HTML of the corresponding recipe.

MEAL="Pizza" && echo "# \n>  | \n<a href=\"\"><img align=\"center\" width=1280 src=\"\"></a>\n\n## Ingredients\n\n| Ingredient | Measure |\n|---|---|\n|||\n\n\n## Instructions\n\n\n\n}" > _template.hbs && oafp url="https://www.themealdb.com/api/json/v1/1/search.php?s=$MEAL" path="set(meals[0],'o').set(k2a(get('o'),'strIngredient','i',\`true\`),'is').set(k2a(get('o'),'strMeasure','m',\`true\`),'ms')|get('o').insert(get('o'),'ingredients',ranges(length(get('is')),\`0\`,\`1\`).map(&{ ingredient: geta('is',@).i, measure: geta('ms',@).m }, @))" out=json | oafp out=template template=_template.hbs | oafp in=md out=html htmlcompact=true

101

πŸ“– Generic | Text

Get a json with lyrics of a song.

curl -s https://api.lyrics.ovh/v1/Coldplay/Viva%20La%20Vida | oafp path="substring(lyrics,index_of(lyrics, '\n'),length(lyrics))"

102

πŸ“– Generic | Text

Search a word in the English dictionary returning phonetic, meanings, synonyms, antonyms, etc.

WORD="google" && oafp url="https://api.dictionaryapi.dev/api/v2/entries/en/$WORD"

103

πŸ“– Generic | URL

Given an URL to a resource on a website determine how long ago is was modified given the data provided by the server.

URL="https://openaf.io/openaf.jar" && oafp url="$URL" urlmethod=head path="response.timeago(to_ms(\"last-modified\"))"

104

πŸ“– Generic | YAML

Given an YAML file with a data array composed of maps with fields 'c', 's', 'd' and 'e' filter by any record where any field doesn't have contents.

oafp oafp-examples.yaml path="data[?length(nvl(c, \'\')) == \`0\` || length(nvl(s, \'\')) == \`0\` || length(nvl(d, \'\')) == \`0\` || length(nvl(e, \'\')) == \`0\`]"

105

πŸ“– GitHub | GIST

Using GitHub's GIST functionality retrieves and parses an oAFp examples YAML file with the template and the corresponding data.

opack install GIST
oafp libs="@GIST/gist.js" in=ch inch="(type: gist)" data="(id: '557e12e4a840d513635b3a57cb57b722', file: oafp-examples.yaml)" path=content | oafp in=yaml out=template templatedata=data templatepath=tmpl

106

πŸ“– GitHub | Releases

Builds a table of GitHub project releases

curl -s https://api.github.com/repos/openaf/openaf/releases | oafp sql="select name, tag_name, published_at order by published_at" output=ctable

107

πŸ“– GitHub | Releases

Parses the latest GitHub project release markdown notes

curl -s https://api.github.com/repos/openaf/openaf/releases | oafp path="[0].body" output=md

108

πŸ“– Grid | Java

Parses a Java hsperf data + the current rss java process memory into a looping grid.

JPID=12345 && HSPERF=/tmp/hsperfdata_openvscode-server/$JPID && oafp in=oafp data="[(file: $HSPERF, in: hsperf, path: java)|(cmd: 'ps -p $JPID -o rss=', path: '{ rss: mul(@,\`1024\`) }')]" merge=true out=grid grid="[[(title:Threads,type:chart,obj:'int threads.live:green:live threads.livePeak:red:peak threads.daemon:blue:daemon -min:0')|(title:RSS,type:chart,obj:'bytes rss:blue:rss')]|[(title:Heap,type:chart,obj:'bytes __mem.total:red:total __mem.used:blue:used -min:0')|(title:Metaspace,type:chart,obj:'bytes __mem.metaTotal:blue:total __mem.metaUsed:green:used -min:0')]]" loop=1

109

πŸ“– Grid | Java

Parses a Java hsperf data into a looping grid.

HSPERF=/tmp/hsperfdata_user/12345 && oafp $HSPERF in=hsperf path=java out=grid grid="[[(title:Threads,type:chart,obj:'int threads.live:green:live threads.livePeak:red:peak threads.daemon:blue:daemon -min:0')|(title:Class Loaders,type:chart,obj:'int cls.loadedClasses:blue:loaded cls.unloadedClasses:red:unloaded')]|[(title:Heap,type:chart,obj:'bytes __mem.total:red:total __mem.used:blue:used -min:0')|(title:Metaspace,type:chart,obj:'bytes __mem.metaTotal:blue:total __mem.metaUsed:green:used -min:0')]]" loop=1

110

πŸ“– Grid | Kubernetes

Displays a continuous updating grid with a line chart with the number of CPU throtlles and bursts recorded in the Linux cgroup cpu stats of a container running in Kubernetes and the source cpu.stats data

oafp cmd="cat /sys/fs/cgroup/cpu.stat | sed 's/ /: /g'" in=yaml out=grid grid="[[(title:cpu.stat,type:tree)|(title:chart,type:chart,obj:'int nr_throttled:red:throttled nr_bursts:blue:bursts -min:0 -vsize:10')]]" loop=1

111

πŸ“– Grid | Mac

Shows a grid with the Mac network metrics and 4 charts for in, out packets and in, out bytes

# opack install mac
sudo powermetrics --help > /dev/null
oafp libs=Mac cmd="sudo powermetrics --format=plist --show-initial-usage -n 0 --samplers network" in=plist path=network out=grid grid="[[(title:data,path:@,xsnap:2)]|[(title:in packets,type:chart,obj:'int ipackets:blue:in')|(title:out packets,type:chart,obj:'int opackets:red:out')]|[(title:in bytes,type:chart,obj:'int ibytes:blue:in')|(title:out bytes,type:chart,obj:'int obytes:red:out')]]" loop=1

112

πŸ“– Grid | Mac

Shows a grid with the Mac storage metrics and 4 charts for read, write IOPS and read, write bytes per second

# opack install mac
sudo powermetrics --help > /dev/null
oafp libs=Mac cmd="sudo powermetrics --format=plist --show-initial-usage -n 0 --samplers disk" in=plist path=disk out=grid grid="[[(title:data,path:@,xsnap:2)]|[(title:read iops,type:chart,obj:'dec3 rops_per_s:blue:read_iops')|(title:write iops,type:chart,obj:'dec3 wops_per_s:red:write_iops')]|[(title:read bytes per sec,type:chart,obj:'bytes rbytes_per_s:blue:read_bytes_per_sec')|(title:write bytes per sec,type:chart,obj:'bytes wbytes_per_s:red:write_bytes_per_sec')]]" loop=1

113

πŸ“– Grid | Unix

On an Unix/Linux system supporting 'ps' output formats %cpu and %mem, will output a chart with the percentage of cpu and memory usage of a provided pid (e.g. 12345)

oafp cmd="ps -p 12345 -o %cpu,%mem" in=lines linesvisual=true linesvisualsepre="\\s+" out=chart chart="int '\"%CPU\"':red:cpu '\"%MEM\"':blue:mem -min:0 -max:100" loop=1 loopcls=true

114

πŸ“– JSON Schemas | Lists

Get a list of JSON schemas from Schema Store catalog

oafp cmd="curl https://raw.githubusercontent.com/SchemaStore/schemastore/master/src/api/json/catalog.json" path="schemas[].{name:name,description:description,files:to_string(fileMatch)}" out=ctable

115

πŸ“– Java | Certificates

Given a Java keystore will obtain a list of certificates and output them order by the ones that will expire first.

# execute 'ojob ojob.io/java/certs -jobhelp' to get more options
ojob ojob.io/java/certs op=list keystore=mycerts -json | oafp out=ctable sql="select * order by notAfter"

116

πŸ“– Kubernetes | Base64

Given a Kubernetes Config Map or Secret with binary data, retrieves it and stores it locally in a binary file.

FILE=config.zip && kubectl get cm my-configs -n kube-system -o json | oafp path="binaryData.\"$FILE\"" | base64 -d > $FILE

117

πŸ“– Kubernetes | Containers

Parse the Linux cgroup cpu stats on a container running in Kubernetes

cat /sys/fs/cgroup/cpu.stat | sed 's/ /: /g' | oafp in=yaml out=ctree

118

πŸ“– Kubernetes | Helm

Given an Helm release name and the corresponding namespace will produce a table with the timestamps when the corresponding Helm chart hooks have started and completed for the lastest execution and the corresponding phase.

RELEASE=myrelease && NS=my-namespace && oafp cmd="helm status $RELEASE -n $NS -o json" out=ctable path="hooks[].{hookName:name,started_at:to_date(last_run.started_at),completed_at:to_date(last_run.completed_at),elapsed:from_ms(sub(to_number(to_ms(to_date(last_run.completed_at))),to_number(to_ms(to_date(last_run.started_at)))),''),phase:last_run.phase}"

119

πŸ“– Kubernetes | Kubectl

Build a table of the images 'cached' in all Kubernetes nodes using Kubectl and, additionally, provide a summary of the total size per node.

# Master table with all nodes
oafp cmd="kubectl get nodes -o json" path="items[].amerge(status.images,{node:metadata.name})[].{node:node,image:names[1],sizeBytes:sizeBytes}" 
# Summary table of total images size per node
oafp cmd="kubectl get nodes -o json" path="items[].amerge(status.images,{node:metadata.name})[].{node:node,image:names[1],sizeBytes:sizeBytes}" sql='select "node", sum("sizeBytes") "totalSize" group by "node"' out=json | oafp field2byte=totalSize out=ctable

120

πŸ“– Kubernetes | Kubectl

Build an output table with Kubernetes pods with namespace, pod name, container name and corresponding resources using kubectl

kubectl get pods -A -o json | oafp path="items[].amerge({ ns: metadata.namespace, pod: metadata.name, phase: status.phase }, spec.containers[].{ container: name, resources: to_slon(resources) })[]" sql="select ns, pod, container, phase, resources order by ns, pod, container" out=ctable

121

πŸ“– Kubernetes | Kubectl

Build an output table with Kubernetes pods with node, namespace, pod name, container name and corresponding resources using kubectl

kubectl get pods -A -o json | oafp path="items[].amerge({ node: spec.nodeName, ns: metadata.namespace, pod: metadata.name, phase: status.phase }, spec.containers[].{ container: name, resources: to_slon(resources) })[]" sql="select node, ns, pod, container, phase, resources order by node, ns, pod, container" out=ctable

122

πŸ“– Kubernetes | Kubectl

Executes a recursive file list find command in a specific pod, namespace and path converting the result into a table.

NS=default && POD=my-pod-5c9cfb87d4-r6dlp && LSPATH=/data && kubectl exec -n $NS $POD -- find $LSPATH -exec stat -c '{"t":"%F", "p": "%n", "s": %s, "m": "%Y", "e": "%A", "u": "%U", "g": "%G"}' {} \; | oafp in=ndjson ndjsonjoin=true path="[].{type:t,permissions:e,user:u,group:g,size:s,modifiedDate:to_date(mul(to_number(m),\`1000\`)),filepath:p}" from="sort(type,filepath)" out=ctable

123

πŸ“– Kubernetes | Kubectl

Given the list of all Kubernetes objects will produce a list of objects per namespace, kind, apiVersiom, creation timestamp, name and owner.

oafp cmd="kubectl get all -A -o json" path="items[].{ns:metadata.namespace,kind:kind,apiVersion:apiVersion,createDate:metadata.creationTimestamp,name:metadata.name,owner:metadata.join(',',map(&concat(kind, concat('/', name)), nvl(ownerReferences,from_json('[]'))))}" sql="select * order by ns, kind, name" out=ctable

124

πŸ“– Kubernetes | Kubectl

List of Kubernetes CPU, memory and storage stats per node using kubectl

oafp cmd="kubectl get nodes -o json" path="items[].{node:metadata.name,totalCPU:status.capacity.cpu,allocCPU:status.allocatable.cpu,totalMem:to_bytesAbbr(from_bytesAbbr(status.capacity.memory)),allocMem:to_bytesAbbr(from_bytesAbbr(status.allocatable.memory)),totalStorage:to_bytesAbbr(from_bytesAbbr(status.capacity.\"ephemeral-storage\")),allocStorage:to_bytesAbbr(to_number(status.allocatable.\"ephemeral-storage\")),conditions:join(\`, \`,status.conditions[].reason)}" output=ctable

125

πŸ“– Kubernetes | Kubectl

List of Kubernetes pods per namespace and kind using kubectl

oafp cmd="kubectl get pods -A -o json" path="items[].{ns:metadata.namespace,kind:metadata.ownerReferences[].kind,name:metadata.name,status:status.phase,restarts:sum(nvl(status.containerStatuses[].restartCount,from_slon('[0]'))),node:spec.nodeName,age:timeago(nvl(status.startTime,now(\`0\`)))}" sql="select * order by status,name" output=ctable

126

πŸ“– Kubernetes | Kubectl

Produces a list of pods' containers per namespace with the corresponding images and assigned nodes.

kubectl get pods -A -o json | oafp path="items[].amerge({namespace: metadata.namespace, pod: metadata.name, nodeName: spec.nodeName},spec.containers[].{container: name, image: image, pullPolicy: imagePullPolicy})[]" sql="select namespace, pod, container, image, pullPolicy, nodeName order by namespace, pod, container" out=ctable

127

πŸ“– Kubernetes | Kubectl

Using kubectl with the appropriate permissions check the filesystem available, capacity and used bytes and inodes on each node of the Kubernetes cluster.

oafp cmd="kubectl get nodes -o json" in=json path="items[].metadata.name" out=cmd outcmd="kubectl get --raw '/api/v1/nodes/{}/proxy/stats/summary' | oafp out=json path=\"node.fs.insert(@,'node','{}')\"" outcmdparam=true | oafp in=ndjson ndjsonjoin=true  out=ctable

128

πŸ“– Kubernetes | PVC

Produces a table with all Kubernetes persistent volume claims (PVCs) in use by pods.

oafp cmd="kubectl get pods -A -o json" path="items[].spec.set(@,'m').volumes[?persistentVolumeClaim].insert(@,'pname',get('m').containers[0].name).insert(@,'node',get('m').nodeName) | [].{pod:pname,node:node,pvc:persistentVolumeClaim.claimName}" from="sort(node,pod)" out=ctable

129

πŸ“– Mac | Activity

Uses the Mac terminal command 'last' output to build an activity table with user, tty, from, login-time and logout-time

oafp cmd="last --libxo json" path="\"last-information\".last" out=ctable

130

πŸ“– Mac | Brew

List all the packages and corresponding versions installed in a Mac by brew.

brew list --versions | oafp in=lines linesjoin=true path="[].split(@,' ').{package:[0],version:[1]}|sort_by(@,&package)" out=ctable

131

πŸ“– Mac | Chart

On a Mac OS produce a looping chart with the total percentage of current CPU usage.

oafp cmd="top -l 1 | grep 'CPU usage' | awk '{print \$3 + \$5}'" out=chart chart="int @:blue:CPU_Usage -min:0 -max:100" loop=2 loopcls=true

132

πŸ“– Mac | Info

Get a list of the current logged users in Mac OS

oafp cmd="who -aH" in=lines linesvisual=true linesjoin=true out=ctable path="[0:-1]"

133

πŸ“– Mac | Info

Parses the current Mac OS hardware information

system_profiler SPHardwareDataType -json | oafp path="SPHardwareDataType[0]" out=ctree

134

πŸ“– Mac | Info

Parses the current Mac OS overview information

system_profiler SPSoftwareDataType -json | oafp path="SPSoftwareDataType[0]" out=ctree

135

πŸ“– Mac | Safari

Get a list of all Mac OS Safari bookmarks into a CSV file.

# opack install mac
oafp ~/Library/Safari/Bookmarks.plist libs=Mac path="Children[].map(&{category:get('cat'),title:URIDictionary.title,url:URLString},setp(@,'Title','cat').nvl(Children,from_json('[]')))[][]" out=csv > bookmarks.csv

136

In a Mac OS with Tunnelblink, if you want to copy all your OpenVPN configurations into ovpn files.

oafp in=ls data="$HOME/Library/Application Support/Tunnelblick/Configurations" path="[?filename=='config.ovpn'].insert(@,'name',replace(filepath,'.+\/([^\/]+)\.tblk\/.+','','\$1'))" lsrecursive=true out=cmd outcmdtmpl=true outcmd="cp \"\" output/\".ovpn\""

137

πŸ“– Markdown | Tables

For an input markdown file, parse all tables, transform it to JSON and output as a colored table

oafp url="https://raw.githubusercontent.com/OpenAF/sh/refs/heads/main/README.md" in=mdtable inmdtablejoin=true path="[0]" out=ctable

138

πŸ“– Network | ASN

Retrieve an IP to ASN list list and converts it to ndjson

oafp cmd="curl https://api.iptoasn.com/data/ip2asn-combined.tsv.gz | gunzip" in=lines linesjoin=true path="[?length(@)>'0'].split(@,'\t').{start:[0],end:[1],asn:[2],area:[3],name:[4]}" out=ndjson

139

πŸ“– Network | ASN

Retrieve the list of ASN number and names from RIPE and transforms it to a CSV.

oafp url="https://ftp.ripe.net/ripe/asnames/asn.txt" in=lines linesjoin=true path="[?length(@)>'0'].split(@,' ').{asn:[0],name:join(' ',[1:])}" out=csv

140

πŸ“– Network | Latency

Given a host and a port will display a continuously updating line chart with network latency, in ms, between the current device and the target host and port

HOST=1.1.1.1 && PORT=53 && oafp in=oaf data="ow.loadNet().testPortLatency('$HOST',$PORT)" out=chart chart="int @:red:latencyMS -min:0" loop=1 loopcls=true

141

πŸ“– Ollama | List models

Parses the list of models currently in an Ollama deployment

export OAFP_MODEL="(type: ollama, model: 'llama3', url: 'https://models.local', timeout: 900000)"
oafp in=llmmodels data="()" out=ctable path="[].{name:name,parameters:details.parameter_size,quantization:details.quantization_level,format:details.format,family:details.family,parent:details.parent,size:size}" sql="select * order by parent,family,format,parameters,quantization"

142

πŸ“– OpenAF | Channels

Copy the json result of a command into an etcd database using OpenAF's channels

oaf -c "\$o(io.listFiles('.').files,{__format:'json'})" | oafp out=ch ch="(type: etcd3, options: (host: localhost, port: 2379), lib: 'etcd3.js')" chkey=canonicalPath

143

πŸ“– OpenAF | Channels

Getting all data stored in an etcd database using OpenAF's channels

echo "" | oafp in=ch inch="(type: etcd3, options: (host: localhost, port: 2379), lib: 'etcd3.js')" out=ctable

144

πŸ“– OpenAF | Channels

Given a Prometheus database will query for a specific metric (go_memstats_alloc_bytes), during a defined period, every 5 seconds (step) will produce a static chart with the corresponding metric values.

URL="http://localhost:9090" && METRIC="go_memstats_alloc_bytes" && TYPE="bytes" && LABELS="job=\"prometheus\"" && START="2024-06-18T20:00:00Z" && END="2024-06-18T20:15:00Z" && STEP=5 && echo "{query:'max($METRIC{$LABELS})',start:'$START',end:'$END',step:$STEP}" | oafp in=ch inch="(type:prometheus,options:(urlQuery:'$URL'))" inchall=true out=json | oafp path="[].set(@, 'main').map(&{metric:'$METRIC',job:get('main').metric.job,timestamp:to_date(mul([0],\`1000\`)),value:to_number([1])}, values) | []" out=schart schart="$TYPE '[].value':green:$METRIC -min:0"

145

πŸ“– OpenAF | Channels

Perform a query to a metric & label, with a start and end time, to a Prometheus server using OpenAF's channels

oafp in=ch inch="(type:prometheus,options:(urlQuery:'http://prometheus.local'))" inchall=true data="(start:'2024-03-22T19:00:00.000Z',end:'2024-03-22T19:05:00.000Z',step:60,query:go_memstats_alloc_bytes_total{job=\"prometheus\"})" path="[].values[].{date:to_date(mul([0],to_number('1000'))),value:[1]}" out=ctable

146

πŸ“– OpenAF | Channels

Retrieve all keys stores in a H2 MVStore file using OpenAF's channels

echo "" | oafp in=ch inch="(type: mvs, options: (file: data.db))" out=ctable

147

πŸ“– OpenAF | Channels

Store and retrieve data from a Redis database

# Install rocksdb opack: 'opack install redis'
#
# Storing data
oafp cmd="oaf -c \"sprint(listFilesRecursive('/usr/bin'))\"" out=ch ch="(type: redis, lib: redis.js, options: (host: '127.0.0.1', port: 6379))" chkey=canonicalPath
# Retrieve data
echo "" | oafp in=ch inch="(type: redis, lib: redis.js, options: (host: '127.0.0.1', port: 6379))" out=pjson

148

πŸ“– OpenAF | Channels

Store and retrieve data from a RocksDB database

# Install rocksdb opack: 'opack install rocksdb'
#
# Storing data
oafp cmd="oaf -c \"sprint(listFilesRecursive('/usr/bin'))\"" out=ch ch="(type: rocksdb, lib: rocksdb.js, options: (path: db))" chkey=canonicalPath
# Retrieve data
echo "" | oafp in=ch inch="(type: rocksdb, lib: rocksdb.js, options: (path: db))" out=pjson

149

πŸ“– OpenAF | Channels

Store the json results of a command into a H2 MVStore file using OpenAF's channels

oaf -c "\$o(listFilesRecursive('.'),{__format:'json'})" | oafp out=ch ch="(type: mvs, options: (file: data.db))" chkey=canonicalPath

150

πŸ“– OpenAF | Flags

List the current values of OpenAF/oAFp internal flags

oafp in=oaf data="__flags"

151

πŸ“– OpenAF | Network

Gets all the DNS host addresses for a provided domain and ensures that the output is always a list

DOMAIN="nattrmon.io" && oafp in=oaf data="ow.loadNet().getDNS('$DOMAIN','a',__,true)" path="if(type(@)=='array',[].Address.HostAddress,[Address.HostAddress]).map(&{ip:@},[])" out=ctable

152

πŸ“– OpenAF | Network

List all MX (mail servers) network addresses from the current DNS server for a hostname using OpenAF

DOMAIN=gmail.com && TYPE=MX && oaf -c "sprint(ow.loadNet().getDNS('$DOMAIN','$TYPE'))" | oafp from="sort(Address)" out=ctable

153

πŸ“– OpenAF | Network

List all network addresses returned from the current DNS server for a hostname using OpenAF

DOMAIN=yahoo.com && oaf -c "sprint(ow.loadNet().getDNS('$DOMAIN'))" | oafp from="sort(Address)" out=ctable

154

πŸ“– OpenAF | OS

Current OS information visible to OpenAF

oafp -v path=os

155

πŸ“– OpenAF | OS

Using OpenAF parse the current environment variables

oaf -c "sprint(getEnvs())" | oafp sortmapkeys=true out=ctree

156

πŸ“– OpenAF | OpenVPN

Using OpenAF code to perform a more complex parsing of the OpenVPN status data running on an OpenVPN container (nmaguiar/openvpn) called 'openvpn'

oafp in=oaf data='(function(){return(b=>{var a=b.split("\n"),c=a.indexOf("ROUTING TABLE"),d=a.indexOf("GLOBAL STATS"),f=a.indexOf("END");b=$csv().fromInString($path(a,`[2:${c}]`).join("\n")).toOutArray();var g=$csv().fromInString($path(a,`[${c+1}:${d}]`).join("\n")).toOutArray();a=$csv().fromInString($path(a,`[${d+1}:${f}]`).join("\n")).toOutArray();return{list:b.map(e=>merge(e,$from(g).equals("Common Name",e["Common Name"]).at(0))),stats:a}})($sh("docker exec openvpn cat /tmp/openvpn-status.log").get(0).stdout)})()' path="list[].{user:\"Common Name\",ip:split(\"Real Address\",':')[0],port:split(\"Real Address\",':')[1],vpnAddress:\"Virtual Address\",bytesRx:to_bytesAbbr(to_number(\"Bytes Received\")),bytesTx:to_bytesAbbr(to_number(\"Bytes Sent\")),connectedSince:to_datef(\"Connected Since\",'yyyy-MM-dd HH:mm:ss'),lastReference:to_datef(\"Last Ref\",'yyyy-MM-dd HH:mm:ss')}" sql="select * order by lastReference" out=ctable

157

πŸ“– OpenAF | SFTP

Generates a file list with filepath, size, permissions, create and last modified time from a SFTP connection with user and password

HOST="my.server" && PORT=22 && LOGIN="user" && PASS=$"abc123" && LSPATH="." && oaf -c "sprint(\$ssh({host:'$HOST',login:'$LOGIN',pass:'$PASS'}).listFiles('$LSPATH'))" | oafp out=ctable path="[].{isDirectory:isDirectory,filepath:filepath,size:size,createTime:to_date(mul(createTime,\`1000\`)),lastModified:to_date(mul(lastModified,\`1000\`)),permissions:permissions}" from="sort(-isDirectory,filepath)"

158

πŸ“– OpenAF | SFTP

Generates a file list with filepath, size, permissions, create and last modified time from a SFTP connection with user, private key and password

HOST="my.server" && PORT=22 && PRIVID=".ssh/id_rsa" && LOGIN="user" && PASS=$"abc123" && LSPATH="." && oaf -c "sprint(\$ssh({host:'$HOST',login:'$LOGIN',pass:'$PASS',id:'$PRIVID'}).listFiles('$LSPATH'))" | oafp out=ctable path="[].{isDirectory:isDirectory,filepath:filepath,size:size,createTime:to_date(mul(createTime,\`1000\`)),lastModified:to_date(mul(lastModified,\`1000\`)),permissions:permissions}" from="sort(-isDirectory,filepath)"

159

πŸ“– OpenAF | TLS

List the TLS certificates of a target host with a sorted alternative names using OpenAF

DOMAIN=yahoo.com && oaf -c "sprint(ow.loadNet().getTLSCertificates('$DOMAIN',443))" | oafp path="[].{issuer:issuerDN,subject:subjectDN,notBefore:notBefore,notAfter:notAfter,alternatives:join(' | ',sort(map(&[1],nvl(alternatives,\`[]\`))))}" out=ctree

160

πŸ“– OpenAF | oJob.io

Parses ojob.io/news results into a clickable news title HMTL page.

ojob ojob.io/news/awsnews __format=json | oafp path="[].{title:replace(t(@,'[]()'),'\|','g','\\|'),date:date}" from="sort(-date)" out=mdtable | oafp in=md out=html

161

πŸ“– OpenAF | oJob.io

Retrieves the list of oJob.io's jobs and filters which start by 'ojob.io/news' to display them in a rectangle

oafp url="https://ojob.io/index.json" path="sort(init.l)[].replace(@,'^https://(.+)\.(yaml|json)$','','\$1')|[?starts_with(@, 'ojob.io/news')]" out=map

162

πŸ“– OpenAF | oPacks

Given a folder of expanded oPacks folders will process each folder .package.yaml file and join each corresponding oPack name and dependencies into a sinlge output map.

oafp in=oafp data="(in:ls,data:.,path:'[?isDirectory].concat(canonicalPath,\`/.package.yaml\`)')" out=cmd outcmd="oafp {} out=json" outcmdparam=true | oafp in=ndjson ndjsonjoin=true path="[].{name:name,deps:dependencies}"

163

πŸ“– OpenAF | oPacks

Listing all currently accessible OpenAF's oPacks

oaf -c "sprint(getOPackRemoteDB())" | oafp maptoarray=true opath="[].{name:name,description:description,version:version}" from="sort(name)" out=ctable

164

πŸ“– OpenAF | oafp

Filter the OpenAF's oafp examples list by a specific word in the description

oafp url="https://ojob.io/oafp-examples.yaml" in=yaml out=template path=data templatepath=tmpl sql="select * where d like '%something%'"

165

πŸ“– OpenAF | oafp

List the OpenAF's oafp examples by category, sub-category and description

oafp url="https://ojob.io/oafp-examples.yaml" in=yaml path="data[].{category:c,subCategory:s,description:d}" from="sort(category,subCategory,description)" out=ctable

166

πŸ“– OpenAF | oafp

Produce a colored table with all the current oafp input and output formats supported.

oafp -v path="concat(oafp.inputs[].{option:'in',type:@}, oafp.outputs[].{option:'out',type:@})" out=ctable

167

πŸ“– OpenVPN | List

When using the container nmaguiar/openvpn it's possible to convert the list of all clients order by expiration/end date

oafp cmd="docker exec openvpn ovpn_listclients" in=csv path="[].{name:name,begin:to_datef(begin,'MMM dd HH:mm:ss yyyy z'),end:to_datef(end,'MMM dd HH:mm:ss yyyy z'),status:status}" out=ctable sql="select * order by end desc"

168

πŸ“– QR | Encode JSON

Given a JSON input encode and decote it from a QR-code png file.

# opack install qr
# Encoding an input into qr.png
oafp in=ls data="." out=json sql="select filepath, size limit 20" | oafp libs=qr in=raw out=qr qrfile=qr.png
# Decoding a qr.png back to json
oafp libs=qr in=qr data=qr.png out=raw | oafp in=json out=ctable

169

πŸ“– QR | Read QR-code

Given a QR-code png file output the corresponding contents.

# opack install qr
oafp libs=qr in=qr data=qr.png out=raw

170

πŸ“– QR | URL

Generate a QR-code for a provided URL.

# opack install qr
oafp libs=qr in=raw data="https://oafp.io" out=qr qrfile=qr.png

171

πŸ“– Unix | Activity

Uses the Linux command 'last' output to build a table with user, tty, from and period of activity for Debian based Linuxs

oafp cmd="last" in=lines linesjoin=true path="[:-3]|[?contains(@,'no logout')==\`false\`&&contains(@,'system boot')==\`false\`].split_re(@,' \\s+').{user:[0],tty:[1],from:[2],period:join(' ',[3:])}" out=ctable

172

πŸ“– Unix | Activity

Uses the Linux command 'last' output to build a table with user, tty, from and period of activity for RedHat based Linuxs

last | sed '/^$/d;$d;$d' | oafp in=lines linesjoin=true path="[].split_re(@, '\\s+').{user: [0], tty: [1], from: [2], login_time: join(' ', [3:7])}" out=ctable

173

πŸ“– Unix | Alpine

List all installed packages in an Alpine system

apk list -I | oafp in=lines linesjoin=true path="[].replace(@,'(.+) (.+) {(.+)} \((.+)\) \[(.+)\]','','\$1|\$2|\$3|\$4').split(@,'|').{package:[0],arch:[1],source:[2],license:[3]}" out=ctable

174

πŸ“– Unix | Ask

Unix bash script to ask for a path and choose between filetypes to perform an unix find command.

#!/bin/bash
temp_file=$(mktemp)
oafp data="[\
  (name: path, prompt: 'Enter the path to search: ', type: question) |\
  (name: type, prompt: 'Choose the file type: ', type: choose, options: ['.jar'|'.class'|'.java'|'.js']) ]"\
     in=ask out=envs\
     arraytomap=true arraytomapkey=name envsnoprefix=true\
     outfile=$temp_file
source $temp_file

find "${path_answer:-.}" -type f -name "*$type_answer"

175

πŸ“– Unix | Compute

Parses the Linux /proc/cpuinfo into an array

cat /proc/cpuinfo | sed "s/^$/---/mg" | oafp in=yaml path="[?not_null(@)]|[?type(processor)=='number']" out=ctree

176

πŸ“– Unix | Debian/Ubuntu

List all installed packages in a Debian/Ubuntu system

apt list --installed | sed "1d" | oafp in=lines linesjoin=true path="[].split(@,' ').{pack:split([0],'/')[0],version:[1],arch:[2]}" out=ctable

177

πŸ“– Unix | Envs

Converts the Linux envs command result into a table of environment variables and corresponding values

env | oafp in=ini path="map(&{key:@,value:to_string(get(@))},sort(keys(@)))" out=ctable

178

πŸ“– Unix | Files

Converting the Linux's /etc/os-release to SQL insert statements.

oafp cmd="cat /etc/os-release" in=ini outkey=release path="[@]" sql="select '$HOSTNAME' \"HOST\", *" out=sql sqlnocreate=true

179

πŸ“– Unix | Files

Converting the Unix's syslog into a json output.

cat syslog | oafp in=raw path="split(trim(@),'\n').map(&split(@, ' ').{ date: concat([0],concat(' ',[1])), time: [2], host: [3], process: [4], message: join(' ',[5:]) }, [])"

180

πŸ“– Unix | Files

Executes a recursive file list find command converting the result into a table.

LSPATH=/openaf && find $LSPATH -exec stat -c '{"t":"%F", "p": "%n", "s": %s, "m": "%Y", "e": "%A", "u": "%U", "g": "%G"}' {} \; | oafp in=ndjson ndjsonjoin=true path="[].{type:t,permissions:e,user:u,group:g,size:s,modifiedDate:to_date(mul(to_number(m),\`1000\`)),filepath:p}" from="sort(type,filepath)" out=ctable

181

πŸ“– Unix | Files

Parses the Linux /etc/passwd to a table order by uid and gid.

oafp cmd="cat /etc/passwd" in=csv inputcsv="(withHeader: false, withDelimiter: ':')" path="[].{user:f0,pass:f1,uid:to_number(f2),gid:to_number(f3),description:f4,home:f5,shell:f6}" out=json | oafp from="notStarts(user, '#').sort(uid, gid)" out=ctable

182

πŸ“– Unix | Generic

Creates, in unix, a data.ndjson file where each record is formatted from json files in /some/data

find /some/data -name "*.json" -exec oafp {} output=json \; > data.ndjson

183

πŸ“– Unix | Memory map

Given an Unix process will output a table with process's components memory address, size in bytes, permissions and owner

pmap 12345 | sed '1d;$d' | oafp in=lines linesjoin=true path="[].split_re(@, '\\s+').{address:[0],size:from_bytesAbbr([1]),perm:[2],owner:join('',[3:])}" out=ctable

184

πŸ“– Unix | Network

Loop over the current Linux active network connections

oafp cmd="netstat -tun | sed \"1d\"" in=lines linesvisual=true linesjoin=true linesvisualsepre="\\s+(\\?\!Address)" out=ctable loop=1

185

πŸ“– Unix | Network

Parse the Linux 'arp' command output

arp | oafp in=lines linesvisual=true linesjoin=true out=ctable

186

πŸ“– Unix | Network

Parse the Linux 'ip tcp_metrics' command

ip tcp_metrics | sed 's/^/target: /g' | sed 's/$/\n\n---\n/g' | sed 's/ \([a-z]*\) /\n\1: /g' | head -n -2 | oafp in=yaml path="[].{target:target,age:from_timeAbbr(replace(age,'[sec|\.]','','')),cwnd:cwnd,rtt:from_timeAbbr(rtt),rttvar:from_timeAbbr(rttvar),source:source}" sql="select * order by target" out=ctable

187

πŸ“– Unix | Network

Parse the result of the Linux route command

route | sed "1d" | oafp in=lines linesjoin=true linesvisual=true linesvisualsepre="\s+" out=ctable

188

πŸ“– Unix | OpenSuse

List all installed packages in an OpenSuse system or zypper based system

zypper se -is | egrep "^i" | oafp in=lines linesjoin=true path="[].split(@,'|').{name:[1],version:[2],arch:[3],repo:[4]}" out=ctable

189

πŸ“– Unix | RedHat

List all installed packages in a RedHat system or rpm based system (use rpm –querytags to list all fields available)

rpm -qa --qf "%{NAME}|%{VERSION}|%{PACKAGER}|%{VENDOR}|%{ARCH}\n" | oafp in=lines linesjoin=true path="[].split(@,'|').{package:[0],version:[1],packager:[2],vendor:[3],arch:[4]}" from="sort(package)" out=ctable

190

πŸ“– Unix | Storage

Converting the Unix's df output

df --output=target,fstype,size,used,avail,pcent | tail -n +2 | oafp in=lines linesjoin=true path="[].split_re(@, ' +').{filesystem:[0],type:[1],size:[2],used:[3],available:[4],use:[5]}" out=ctable

191

πŸ“– Unix | Storage

Parses the result of the Unix ls command

ls -lad --time-style="+%Y-%m-%d %H:%M" * | oafp in=lines path="map(&split_re(@,'\\s+').{permissions:[0],id:[1],user:[2],group:[3],size:[4],date:[5],time:[6],file:[7]},[])" linesjoin=true out=ctable

192

πŸ“– Unix | SystemCtl

Converting the Unix's systemctl list-timers

systemctl list-timers | head -n -3 | oafp in=lines linesvisual=true linesjoin=true out=ctable

193

πŸ“– Unix | SystemCtl

Converting the Unix's systemctl list-units

systemctl list-units | head -n -6 | oafp in=lines linesvisual=true linesjoin=true path="[].delete(@,'')" out=ctable

194

πŸ“– Unix | SystemCtl

Converting the Unix's systemctl list-units into an overview table

systemctl list-units | head -n -6 | oafp in=lines linesvisual=true linesjoin=true path="[].delete(@,'')" sql="select \"LOAD\", \"ACTIVE SUB\", count(1) as \"COUNT\" group by \"LOAD\", \"ACTIVE SUB\"" sqlfilter=advanced out=ctable

195

πŸ“– Unix | Threads

Given an unix process id (pid) loop a table with its top 25 most cpu active threads

JPID=12345 && oafp cmd="ps -L -p $JPID -o tid,pcpu,comm|tail +2" in=lines linesjoin=true path="[].split_re(trim(@),'\s+').{tid:[0],thread:join(' ',[2:]),cpu:to_number(nvl([1],\`-1\`)),cpuPerc:progress(nvl(to_number([1]),\`0\`), \`100\`, \`0\`, \`50\`, __, __)}" sql='select * order by cpu desc limit 25' out=ctable loop=1 loopcls=true

196

πŸ“– Unix | UBI

List all installed packages in an UBI system

microdnf repoquery --setopt=cachedir=/tmp --installed | oafp in=lines linesjoin=true path="[].replace(@,'(.+)\.(\w+)\.(\w+)\$','','\$1|\$2|\$3').split(@,'|').{package:[0],dist:[1],arch:[2]}" out=ctable

197

πŸ“– Unix | named

Converts a Linux's named log, for client queries, into a CSV

cat named.log | oafp in=lines linesjoin=true path="[?contains(@,' client ')==\`true\`].split(@,' ').{datetime:to_datef(concat([0],concat(' ',[1])),'dd-MMM-yyyy HH:mm:ss.SSS'),session:[3],sourceIP:replace([4],'(.+)#(\d+)','','\$1'),sourcePort:replace([4],'(.+)#(\d+)','','\$2'),target:replace([5],'\((.+)\):','','\$1'),query:join(' ',[6:])}" out=csv

198

πŸ“– Unix | strace

Given a strace unix command will produce a summary table of the system calls invoked including a small line chart of the percentage of time of each.

strace -c -o '!oafp in=lines linesvisual=true linesjoin=true opath="[1:-2].merge(@,{perc:progress(to_number(\"% time\"),\`100\`,\`0\`,\`15\`,__,__)})" out=ctable' strace --tips

199

πŸ“– VSCode | Extensions

Check a Visual Studio Code (vscode) extension (vsix) manifest.

oafp in=xml file="Org.my-extension.vsix::extension.vsixmanifest" out=ctree

200

πŸ“– Windows | Network

Output a table with the current route table using Windows' PowerShell

Get-NetRoute | ConvertTo-Json | .\oafp.bat path="[].{destination:DestinationPrefix,gateway:NextHop,interface:InterfaceAlias,metric:InterfaceMetric}" sql=select\ *\ order\ by\ interface,destination out=ctable

201

πŸ“– Windows | Network

Output a table with the list of network interfaces using Windows' PowerShell

Get-NetIPAddress | ConvertTo-Json | .\oafp.bat path="[].{ipAddress:IPAddress,prefixLen:PrefixLength,interface:InterfaceAlias}" sql=select\ *\ order\ by\ interface out=ctable

202

πŸ“– Windows | PnP

Output a table with USB/PnP devices using Windows' PowerShell

Get-PnpDevice -PresentOnly | ConvertTo-Csv -NoTypeInformation | .\oafp.bat in=csv path="[].{class:PNPClass,service:Service,name:FriendlyName,id:InstanceId,description:Description,deviceId:DeviceID,status:Status,present:Present}" sql=select\ *\ order\ by\ class,service out=ctable

203

πŸ“– Windows | Storage

Output a table with the attached disk information using Windows' PowerShell

Get-Disk | ConvertTo-Csv -NoTypeInformation | .\oafp.bat in=csv path="[].{id:trim(UniqueId),name:FriendlyName,isBoot:IsBoot,location:Location,size:to_bytesAbbr(to_number(Size)),allocSize:to_bytesAbbr(to_number(AllocatedSize)),sectorSize:LogicalSectorSize,phySectorSize:PhysicalSectorSize,numPartitions:NumberOfPartitions,partitioning:PartitionStyle,health:HealthStatus,bus:BusType,manufacturer:Manufacturer,model:Model,firmwareVersion:FirmwareVersion,serialNumber:SerialNumber}" out=ctable

204

πŸ“– XML | Maven

Given a Maven pom.xml parses the XML content to a colored table ordering by the fields groupId and artifactId.

oafp pom.xml path="project.dependencies.dependency" out=ctable sql="select * order by groupId, artifactId"

205

πŸ“– nAttrMon | Plugs

Given a nAttrMon config folder, with YAML files, produce a summary table with the each plug (yaml file) execFrom definition.

oafp cmd="grep -R execFrom" in=lines path="[].split(@,':').{plug:[0],execFrom:[2]}" linesjoin=true out=ctable sql="select execFrom, plug where plug <> '' order by execFrom"