ysfad's picture
Upload README.md with huggingface_hub
c46d32e verified
metadata
library_name: pytorch
tags:
  - waste-classification
  - mae
  - vision-transformer
  - environmental
  - recycling
license: mit
datasets:
  - RealWaste
metrics:
  - accuracy
model-index:
  - name: MAE Waste Classifier
    results:
      - task:
          type: image-classification
          name: Waste Classification
        dataset:
          type: RealWaste
          name: RealWaste Dataset
        metrics:
          - type: accuracy
            value: 0.9327
            name: Validation Accuracy

MAE Waste Classifier

A finetuned MAE (Masked Autoencoder) ViT-Base model for waste classification achieving 93.27% validation accuracy on 9 waste categories.

Model Details

  • Architecture: Vision Transformer (ViT-Base) with MAE pretraining
  • Parameters: ~86M
  • Input Size: 224x224 RGB images
  • Classes: 9 waste categories
  • Validation Accuracy: 93.27%

Categories

  1. Cardboard - Flatten and place in recycling bin. Remove any tape or staples.
  2. Food Organics - Compost in organic waste bin or home composter.
  3. Glass - Rinse and place in glass recycling. Remove lids and caps.
  4. Metal - Rinse aluminum/steel cans and place in recycling bin.
  5. Miscellaneous Trash - Dispose in general waste bin. Cannot be recycled.
  6. Paper - Place clean paper in recycling. Remove plastic windows from envelopes.
  7. Plastic - Check recycling number. Rinse containers before recycling.
  8. Textile Trash - Donate if reusable, otherwise dispose in textile recycling.
  9. Vegetation - Compost in organic waste or use for mulch in garden.

Usage

import torch
import timm
from PIL import Image
from torchvision import transforms

# Load model
model = timm.create_model('vit_base_patch16_224', pretrained=False, num_classes=9)
checkpoint = torch.load('best_model.pth', map_location='cpu')
model.load_state_dict(checkpoint['model_state_dict'])
model.eval()

# Preprocessing
transform = transforms.Compose([
    transforms.Resize((224, 224)),
    transforms.ToTensor(),
    transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
])

# Inference
image = Image.open('waste_item.jpg').convert('RGB')
input_tensor = transform(image).unsqueeze(0)

with torch.no_grad():
    outputs = model(input_tensor)
    probabilities = torch.nn.functional.softmax(outputs, dim=1)
    predicted_class = torch.argmax(probabilities, dim=1).item()

categories = ['Cardboard', 'Food Organics', 'Glass', 'Metal', 'Miscellaneous Trash', 'Paper', 'Plastic', 'Textile Trash', 'Vegetation']
print(f"Predicted: {categories[predicted_class]}")

Training Details

  • Dataset: RealWaste (4,752 images)
  • Pretraining: MAE on ImageNet
  • Finetuning: 15 epochs on RealWaste
  • Optimizer: AdamW
  • Hardware: NVIDIA RTX 3080 Ti

Performance

  • Validation Accuracy: 93.27%
  • Training Accuracy: 99.89%
  • Model Size: ~350MB
  • Inference Speed: ~50ms per image (GPU)

Environmental Impact

This model helps improve recycling efficiency by providing accurate waste classification and proper disposal instructions.