PyTorch
llama
File size: 3,127 Bytes
f99e404
 
 
 
 
 
 
a76121d
 
 
 
f99e404
a76121d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
---
license: apache-2.0
metrics:
- accuracy
base_model:
- meta-llama/Llama-3.1-8B-Instruct
---

# Model Card for Llama8b-NNetNav-WA

<!-- Provide a quick summary of what the model is/does. [Optional] -->
LLama8b-NNetNav-WA is a [LLama-3.1-8B]() model that is instruct-tuned with [NNetNav]() data collected via unsupervised exploration on WebArena websites, with a larger [LLama-3.1-70B]() model.

Most details about this model along with details can be found in our paper: [NNetNav: Unsupervised Learning of Browser Agents Through Environment Interaction in the Wild](https://arxiv.org/abs/2410.02907).


![show an example trajectory from NNetNav-WA](TODO)

##  Table of Contents

- [Model Card for Llama8b-NNetNav-WA](#model-card-for--model_id-)
- [Table of Contents](#table-of-contents)
- [Model Details](#model-details)
  - [Model Description](#model-description)
- [Uses](#uses)
- [Bias, Risks, and Limitations](#bias-risks-and-limitations)
- [Training Details](#training-details)
  - [Training Data](#training-data)
  - [Training Procedure](#training-procedure)
- [Environmental Impact](#environmental-impact)
- [Technical Specifications [optional]](#technical-specifications-optional)
  - [Model Architecture and Objective](#model-architecture-and-objective)
  - [Compute Infrastructure](#compute-infrastructure)
    - [Hardware](#hardware)
    - [Software](#software)
- [Citation](#citation)
- [Model Card Authors [optional]](#model-card-authors-optional)
- [Model Card Contact](#model-card-contact)
- [How to Get Started with the Model](#how-to-get-started-with-the-model)


## Model Details

### Model Description

<!-- Provide a longer summary of what this model is/does. -->


## Uses



## Bias, Risks, and Limitations

<!-- This section is meant to convey both technical and sociotechnical limitations. -->

## How to Get Started with the Model

```python


```

## Training Details

### Training Data

<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->

This model was trained on the [NNetnav-WA]() corpus.


### Training Procedure

This model was trained for 2 epochs (roughly 4k gradient steps) with a batch size of 128, and a maximum sequence length of 20000.

### Environmental Impact

- **Hardware Type:** 4 H100 GPUs (80G)
- **Hours used:** Roughly 2 days.
- **Cloud Provider:** Stanford compute.
- **Compute Region:** Stanford energy grid.

### Model Architecture and Objective


### Compute Infrastructure

This model was trained on a slurm cluster.

### Hardware

This model was trained on 4 H100s.

### Software

This model was fine-tuned with [Open-Instruct](https://github.com/allenai/open-instruct/tree/main)

## Citation

**BibTeX:**

```

```


## Model Card Authors [optional]

<!-- This section provides another layer of transparency and accountability. Whose views is this model card representing? How many voices were included in its construction? Etc. -->

Shikhar Murty

## Model Card Contact

smurty@cs.stanford.edu
shikhar.murty@gmail.com