LAVIS
AndroidTacticalAssaultKit-CIV
LAVIS | AndroidTacticalAssaultKit-CIV | |
---|---|---|
18 | 16 | |
8,838 | 807 | |
2.9% | 3.6% | |
6.3 | 3.4 | |
24 days ago | 5 months ago | |
Jupyter Notebook | Java | |
BSD 3-clause "New" or "Revised" License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
LAVIS
- FLaNK AI for 11 March 2024
- FLaNK 04 March 2024
-
[D] Why is most Open Source AI happening outside the USA?
For multimodal, there's China (*many), then Salesforce.
-
Need help for a colab notebook running Lavis blip2_instruct_vicuna13b?
Been trying for all day to get a working inference for this example: https://github.com/salesforce/LAVIS/tree/main/projects/instructblip
-
most sane web3 job listing
There's also been big breakthroughs in computer vision. Not that long ago it was hard to recognize if a photo contained a bird; that's solved now by models like CLIP, Yolo, or Segment Anything. Now research has moved on to generating 3D scenes from images or interactively answering questions about images.
-
I work at a non-tech company and have been asked to make software that is impossible. How do I explain it to my boss?
The new hotness is multimodal vision-language models like InstructBLIP that can interactively answer questions about images. Check out the examples in the github repo, I would not have thought this was possible a few years ago.
-
Two-minute Daily AI Update (Date: 5/15/2023)
Salesforce’s BLIP family has a new member– InstructBLIP, a vision-language instruction-tuning framework using BLIP-2 models. It has achieved state-of-the-art zero-shot generalization performance on a wide range of vision-language tasks, substantially outperforming BLIP-2 and Flamingo. (Source)
-
InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning
Github
-
Can I use my own art as a training set?
Most of my workflows are self-made. For captioning I used Blip-2 in a custom script I made that automates the process by going into directories and their sub-directories and creates a .txt file beside each image. This way I can keep my images organized in their proper directories, without having to put dump them all in a single place.
- FLiP Stack Weekly for 13-Feb-2023
AndroidTacticalAssaultKit-CIV
- FLaNK 04 March 2024
-
Cannot find library within SDK
I haven't had that issue before, but this github issues link might help you. Note: I'm new to TAK development so sorry if this isn't helpful. Linky
-
Where is the public SDK for building a Plugin for 4.8.1 (current Playstore version)
I can see the plugin SDK for 4.6 here Releases · deptofdefense/AndroidTacticalAssaultKit-CIV · GitHub but if I use that to build for TAK version 4.8.1 on the 3rd party pipeline, the Plugin won't load into CivTAK 4.8.1, despite being listed as current. I built the plugin with the 4.6 SDK and set ext.ATAK_VERSION = "4.8.1" in the app gradle file.
-
Mesh Radio's
Without access to military gear I know of guys building ATAK https://github.com/deptofdefense/AndroidTacticalAssaultKit-CIV with specialised APIs that support rural Fire services etc . And was thinking that a drone with a FLIR camera that could communicate to the ATAK ground station via a TCP/UDP carrier would be viable. There's also our beach life savers that use drones to surveillance for sharks and rips... I wonder if you had that imagery that could use AI like facial recognition to quickly identify hazards.
-
Civtak Plugin template gradle build error
I'm currently trying to dip my feet into the world of civtak plugin development and have been running into a few issues getting the latest version (SDK version 4.5.1.13) of the sdk to build using gradle on android studio. When gradle tries to build the plugin template project, I keep getting this particular error.
-
'Build an ATAK Dev Environment' Error
I've been working through the available dev guidance to build my first plugin. I started with Ballantyne's 'Build an ATAK Dev Environment' but I haven't been able to get the project to build successfully. I keep getting a Conan install error. Is this a necessary step for ATAK-CIV plugin development?
- Track Management!
-
Never underestimate the power of ISR
Actually, it is open source, since 2020.
-
Does anyone know if there is a schema or explanation somewhere for the .prefs file in a Data Package?
Link to the source code, in case you didn't already have it: https://github.com/deptofdefense/AndroidTacticalAssaultKit-CIV
- (GCC) The customer has nuclear weapons. They do not do “bounty”
What are some alternatives?
pytorch-widedeep - A flexible package for multimodal-deep-learning to combine tabular data with text and images using Wide and Deep models in Pytorch
AndroidTacticalAssaultKit-C
CLIP-Caption-Reward - PyTorch code for "Fine-grained Image Captioning with CLIP Reward" (Findings of NAACL 2022)
ingestr - ingestr is a CLI tool to copy data between any databases with a single command seamlessly.
sparseml - Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
GeneFacePlusPlus - GeneFace++: Generalized and Stable Real-Time 3D Talking Face Generation; Official Code
robo-vln - Pytorch code for ICRA'21 paper: "Hierarchical Cross-Modal Agent for Robotics Vision-and-Language Navigation"
DeepViewAgg - [CVPR'22 Best Paper Finalist] Official PyTorch implementation of the method presented in "Learning Multi-View Aggregation In the Wild for Large-Scale 3D Semantic Segmentation"
linkis - Apache Linkis builds a computation middleware layer to facilitate connection, governance and orchestration between the upper applications and the underlying data engines.
multimodal - A collection of multimodal datasets, and visual features for VQA and captionning in pytorch. Just run "pip install multimodal"
clipseg - This repository contains the code of the CVPR 2022 paper "Image Segmentation Using Text and Image Prompts".
Oscar - Oscar and VinVL