EdjeElectronics
{
"type": "http://schema.org/Person",
"name": "",
"description": "",
"followers": "",
"url": "",
"location": "",
"languages": [
"Python",
"Jupyter Notebook",
"Python",
"Python",
"Jupyter Notebook",
"Python"
],
"users": [
{
"name": "@EdjeElectronics",
"avatar": "https://avatars.githubusercontent.com/u/26154245?s=64&v=4"
},
{
"name": "@EdjeElectronics",
"avatar": "https://avatars.githubusercontent.com/u/26154245?s=64&v=4"
},
{
"name": "View EdjeElectronics's full-sized avatar",
"avatar": "https://avatars.githubusercontent.com/u/26154245?v=4"
}
],
"topics": []
}
EJ Technology Consultants
{
"avatar": "https://avatars.githubusercontent.com/u/26154245?v=4",
"name": "Evan",
"username": "EdjeElectronics",
"description": "Computer vision engineer and founder at EJ Technology Consultants.",
"location": "Bozeman, MT",
"vcard": "<svg class=\"octicon octicon-organization\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" height=\"16\" aria-hidden=\"true\"><path d=\"M1.75 16A1.75 1.75 0 0 1 0 14.25V1.75C0 .784.784 0 1.75 0h8.5C11.216 0 12 .784 12 1.75v12.5c0 .085-.006.168-.018.25h2.268a.25.25 0 0 0 .25-.25V8.285a.25.25 0 0 0-.111-.208l-1.055-.703a.749.749 0 1 1 .832-1.248l1.055.703c.487.325.779.871.779 1.456v5.965A1.75 1.75 0 0 1 14.25 16h-3.5a.766.766 0 0 1-.197-.026c-.099.017-.2.026-.303.026h-3a.75.75 0 0 1-.75-.75V14h-1v1.25a.75.75 0 0 1-.75.75Zm-.25-1.75c0 .138.112.25.25.25H4v-1.25a.75.75 0 0 1 .75-.75h2.5a.75.75 0 0 1 .75.75v1.25h2.25a.25.25 0 0 0 .25-.25V1.75a.25.25 0 0 0-.25-.25h-8.5a.25.25 0 0 0-.25.25ZM3.75 6h.5a.75.75 0 0 1 0 1.5h-.5a.75.75 0 0 1 0-1.5ZM3 3.75A.75.75 0 0 1 3.75 3h.5a.75.75 0 0 1 0 1.5h-.5A.75.75 0 0 1 3 3.75Zm4 3A.75.75 0 0 1 7.75 6h.5a.75.75 0 0 1 0 1.5h-.5A.75.75 0 0 1 7 6.75ZM7.75 3h.5a.75.75 0 0 1 0 1.5h-.5a.75.75 0 0 1 0-1.5ZM3 9.75A.75.75 0 0 1 3.75 9h.5a.75.75 0 0 1 0 1.5h-.5A.75.75 0 0 1 3 9.75ZM7.75 9h.5a.75.75 0 0 1 0 1.5h-.5a.75.75 0 0 1 0-1.5Z\"></path></svg>\n <span class=\"p-org\"><div>EJ Technology Consultants</div></span>\n",
"vcardDetails": [
{
"name": "https://ejtech.io",
"url": "https://ejtech.io"
},
{
"name": "@EdjeElectronics",
"url": "https://twitter.com/EdjeElectronics"
}
],
"orgs": [],
"sponsors": [],
"pinned": [
{
"name": "TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10",
"description": "How to train a TensorFlow Object Detection Classifier for multiple object detection on Windows",
"language": ""
},
{
"name": "TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi",
"description": "A tutorial showing how to train, convert, and run TensorFlow Lite object detection models on Android devices, the Raspberry Pi, and more!",
"language": ""
},
{
"name": "TensorFlow-Object-Detection-on-the-Raspberry-Pi",
"description": "A tutorial showing how to set up TensorFlow's Object Detection API on the Raspberry Pi",
"language": ""
},
{
"name": "OpenCV-Playing-Card-Detector",
"description": "Python program that uses OpenCV to detect and identify playing cards from a PiCamera video feed on a Raspberry Pi",
"language": ""
},
{
"name": "Train-and-Deploy-YOLO-Models",
"description": "Tutorials and examples showing how to train and deploy Ultralytics YOLO models",
"language": ""
},
{
"name": "Image-Dataset-Tools",
"description": "Handy scripts for building and augmenting a machine learning image dataset",
"language": ""
}
],
"pinnedHtml": [
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"120985410\" href=\"/EdjeElectronics/TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10\n </span></a> <tool-tip id=\"tooltip-50bdf601-ecda-465f-b4b5-7116e33cec87\" for=\"120985410\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n How to train a TensorFlow Object Detection Classifier for multiple object detection on Windows\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #3572A5\"></span>\n <span itemprop=\"programmingLanguage\">Python</span>\n</span>\n\n <a href=\"/EdjeElectronics/TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 2.9k\n </a>\n <a href=\"/EdjeElectronics/TensorFlow-Object-Detection-API-Tutorial-Train-Multiple-Objects-Windows-10/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 1.3k\n </a>\n </p>\n ",
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"210062852\" href=\"/EdjeElectronics/TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi\n </span></a> <tool-tip id=\"tooltip-fb170010-f793-4452-ab89-2bfe065187e4\" for=\"210062852\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n A tutorial showing how to train, convert, and run TensorFlow Lite object detection models on Android devices, the Raspberry Pi, and more!\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #DA5B0B\"></span>\n <span itemprop=\"programmingLanguage\">Jupyter Notebook</span>\n</span>\n\n <a href=\"/EdjeElectronics/TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 1.6k\n </a>\n <a href=\"/EdjeElectronics/TensorFlow-Lite-Object-Detection-on-Android-and-Raspberry-Pi/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 697\n </a>\n </p>\n ",
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"136115442\" href=\"/EdjeElectronics/TensorFlow-Object-Detection-on-the-Raspberry-Pi\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n TensorFlow-Object-Detection-on-the-Raspberry-Pi\n </span></a> <tool-tip id=\"tooltip-66e7bdf9-4bf2-4763-9d22-ffeeed12b3d9\" for=\"136115442\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">TensorFlow-Object-Detection-on-the-Raspberry-Pi</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n A tutorial showing how to set up TensorFlow's Object Detection API on the Raspberry Pi\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #3572A5\"></span>\n <span itemprop=\"programmingLanguage\">Python</span>\n</span>\n\n <a href=\"/EdjeElectronics/TensorFlow-Object-Detection-on-the-Raspberry-Pi/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 1.2k\n </a>\n <a href=\"/EdjeElectronics/TensorFlow-Object-Detection-on-the-Raspberry-Pi/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 360\n </a>\n </p>\n ",
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"106424794\" href=\"/EdjeElectronics/OpenCV-Playing-Card-Detector\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n OpenCV-Playing-Card-Detector\n </span></a> <tool-tip id=\"tooltip-5ab3c874-90c3-4bd3-9aa8-5ac5d2c4c522\" for=\"106424794\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">OpenCV-Playing-Card-Detector</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n Python program that uses OpenCV to detect and identify playing cards from a PiCamera video feed on a Raspberry Pi\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #3572A5\"></span>\n <span itemprop=\"programmingLanguage\">Python</span>\n</span>\n\n <a href=\"/EdjeElectronics/OpenCV-Playing-Card-Detector/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 742\n </a>\n <a href=\"/EdjeElectronics/OpenCV-Playing-Card-Detector/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 232\n </a>\n </p>\n ",
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"896120407\" href=\"/EdjeElectronics/Train-and-Deploy-YOLO-Models\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n Train-and-Deploy-YOLO-Models\n </span></a> <tool-tip id=\"tooltip-c5d84c36-5e07-4b12-8f04-aafd6332df22\" for=\"896120407\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">Train-and-Deploy-YOLO-Models</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n Tutorials and examples showing how to train and deploy Ultralytics YOLO models\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #DA5B0B\"></span>\n <span itemprop=\"programmingLanguage\">Jupyter Notebook</span>\n</span>\n\n <a href=\"/EdjeElectronics/Train-and-Deploy-YOLO-Models/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 252\n </a>\n <a href=\"/EdjeElectronics/Train-and-Deploy-YOLO-Models/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 88\n </a>\n </p>\n ",
"\n <div class=\"d-flex v-align-middle mr-2\">\n <span data-view-component=\"true\" class=\"position-relative\"><a id=\"367688053\" href=\"/EdjeElectronics/Image-Dataset-Tools\" data-view-component=\"true\" class=\"min-width-0 Link text-bold flex-auto wb-break-all\"><span class=\"repo\">\n Image-Dataset-Tools\n </span></a> <tool-tip id=\"tooltip-0fb01129-328f-4271-a352-af6c81d0d23e\" for=\"367688053\" popover=\"manual\" data-direction=\"n\" data-type=\"description\" data-view-component=\"true\" class=\"sr-only position-absolute\">Image-Dataset-Tools</tool-tip></span> <span class=\"flex-auto text-right\">\n <span></span><span class=\"Label Label--secondary v-align-middle \">Public</span>\n </span>\n </div>\n\n\n <p class=\"pinned-item-desc color-fg-muted text-small d-block mt-2 tmp-mb-3\">\n Handy scripts for building and augmenting a machine learning image dataset\n </p>\n\n <p class=\"mb-0 f6 color-fg-muted\">\n <span class=\"tmp-mr-3 d-inline-block\">\n <span class=\"repo-language-color\" style=\"background-color: #3572A5\"></span>\n <span itemprop=\"programmingLanguage\">Python</span>\n</span>\n\n <a href=\"/EdjeElectronics/Image-Dataset-Tools/stargazers\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"stars\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-star\">\n <path d=\"M8 .25a.75.75 0 0 1 .673.418l1.882 3.815 4.21.612a.75.75 0 0 1 .416 1.279l-3.046 2.97.719 4.192a.751.751 0 0 1-1.088.791L8 12.347l-3.766 1.98a.75.75 0 0 1-1.088-.79l.72-4.194L.818 6.374a.75.75 0 0 1 .416-1.28l4.21-.611L7.327.668A.75.75 0 0 1 8 .25Zm0 2.445L6.615 5.5a.75.75 0 0 1-.564.41l-3.097.45 2.24 2.184a.75.75 0 0 1 .216.664l-.528 3.084 2.769-1.456a.75.75 0 0 1 .698 0l2.77 1.456-.53-3.084a.75.75 0 0 1 .216-.664l2.24-2.183-3.096-.45a.75.75 0 0 1-.564-.41L8 2.694Z\"></path>\n</svg>\n 41\n </a>\n <a href=\"/EdjeElectronics/Image-Dataset-Tools/forks\" class=\"pinned-item-meta Link--muted\">\n <svg aria-label=\"forks\" role=\"img\" height=\"16\" viewBox=\"0 0 16 16\" version=\"1.1\" width=\"16\" data-view-component=\"true\" class=\"octicon octicon-repo-forked\">\n <path d=\"M5 5.372v.878c0 .414.336.75.75.75h4.5a.75.75 0 0 0 .75-.75v-.878a2.25 2.25 0 1 1 1.5 0v.878a2.25 2.25 0 0 1-2.25 2.25h-1.5v2.128a2.251 2.251 0 1 1-1.5 0V8.5h-1.5A2.25 2.25 0 0 1 3.5 6.25v-.878a2.25 2.25 0 1 1 1.5 0ZM5 3.25a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Zm6.75.75a.75.75 0 1 0 0-1.5.75.75 0 0 0 0 1.5Zm-3 8.75a.75.75 0 1 0-1.5 0 .75.75 0 0 0 1.5 0Z\"></path>\n</svg>\n 14\n </a>\n </p>\n "
]
}
{
"accept-ranges": "bytes",
"cache-control": "max-age=0, private, must-revalidate",
"content-encoding": "gzip",
"content-security-policy": "default-src 'none'; base-uri 'self'; child-src github.githubassets.com github.com/assets-cdn/worker/ github.com/assets/ gist.github.com/assets-cdn/worker/; connect-src 'self' uploads.github.com www.githubstatus.com collector.github.com raw.githubusercontent.com api.github.com github-cloud.s3.amazonaws.com github-production-repository-file-5c1aeb.s3.amazonaws.com github-production-upload-manifest-file-7fdce7.s3.amazonaws.com github-production-user-asset-6210df.s3.amazonaws.com *.rel.tunnels.api.visualstudio.com wss://*.rel.tunnels.api.visualstudio.com github.githubassets.com objects-origin.githubusercontent.com copilot-proxy.githubusercontent.com proxy.individual.githubcopilot.com proxy.business.githubcopilot.com proxy.enterprise.githubcopilot.com *.actions.githubusercontent.com wss://*.actions.githubusercontent.com productionresultssa0.blob.core.windows.net productionresultssa1.blob.core.windows.net productionresultssa2.blob.core.windows.net productionresultssa3.blob.core.windows.net productionresultssa4.blob.core.windows.net productionresultssa5.blob.core.windows.net productionresultssa6.blob.core.windows.net productionresultssa7.blob.core.windows.net productionresultssa8.blob.core.windows.net productionresultssa9.blob.core.windows.net productionresultssa10.blob.core.windows.net productionresultssa11.blob.core.windows.net productionresultssa12.blob.core.windows.net productionresultssa13.blob.core.windows.net productionresultssa14.blob.core.windows.net productionresultssa15.blob.core.windows.net productionresultssa16.blob.core.windows.net productionresultssa17.blob.core.windows.net productionresultssa18.blob.core.windows.net productionresultssa19.blob.core.windows.net github-production-repository-image-32fea6.s3.amazonaws.com github-production-release-asset-2e65be.s3.amazonaws.com insights.github.com wss://alive.github.com wss://alive-staging.github.com api.githubcopilot.com api.individual.githubcopilot.com api.business.githubcopilot.com api.enterprise.githubcopilot.com; font-src github.githubassets.com; form-action 'self' github.com gist.github.com copilot-workspace.githubnext.com objects-origin.githubusercontent.com; frame-ancestors 'none'; frame-src viewscreen.githubusercontent.com notebooks.githubusercontent.com; img-src 'self' data: blob: github.githubassets.com media.githubusercontent.com camo.githubusercontent.com identicons.github.com avatars.githubusercontent.com private-avatars.githubusercontent.com github-cloud.s3.amazonaws.com objects.githubusercontent.com release-assets.githubusercontent.com secured-user-images.githubusercontent.com user-images.githubusercontent.com private-user-images.githubusercontent.com opengraph.githubassets.com marketplace-screenshots.githubusercontent.com copilotprodattachments.blob.core.windows.net/github-production-copilot-attachments/ github-production-user-asset-6210df.s3.amazonaws.com customer-stories-feed.github.com spotlights-feed.github.com objects-origin.githubusercontent.com *.githubusercontent.com; manifest-src 'self'; media-src github.com user-images.githubusercontent.com secured-user-images.githubusercontent.com private-user-images.githubusercontent.com github-production-user-asset-6210df.s3.amazonaws.com gist.github.com github.githubassets.com; script-src github.githubassets.com; style-src 'unsafe-inline' github.githubassets.com; upgrade-insecure-requests; worker-src github.githubassets.com github.com/assets-cdn/worker/ github.com/assets/ gist.github.com/assets-cdn/worker/",
"content-type": "text/html; charset=utf-8",
"date": "Fri, 17 Apr 2026 16:26:29 GMT",
"etag": "604ad092305f8d4c2c29b1f41094c8d9",
"referrer-policy": "origin-when-cross-origin, strict-origin-when-cross-origin",
"server": "github.com",
"set-cookie": "logged_in=no; expires=Sat, 17 Apr 2027 16:26:29 GMT; domain=.github.com; path=/; HttpOnly; secure; SameSite=Lax",
"strict-transport-security": "max-age=31536000; includeSubdomains; preload",
"transfer-encoding": "chunked",
"vary": "X-PJAX, X-PJAX-Container, Turbo-Visit, Turbo-Frame, X-Requested-With, Sec-Fetch-Site,Accept-Encoding, Accept, X-Requested-With",
"x-content-type-options": "nosniff",
"x-frame-options": "deny",
"x-github-request-id": "875A:1DD01A:17C70C1:208368E:69E25F34",
"x-xss-protection": "0"
}