Coder Social home page Coder Social logo

pytorch-yolo-v3's People

Contributors

ayooshkathuria avatar ryoppippi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

pytorch-yolo-v3's Issues

SyntaxError: only named arguments may follow *expression

I tried running your repo on my system running Ubuntu 16.o4, with Python 3.5 and pytorch 0.4. I am running into the following error when I simply try running detect.py

nivedithak@nivii-Sum18:~/Downloads/pytorch-yolo-v3-master$ python detect.py 
  File "detect.py", line 29
    fwd = nn.Sequential(self.linear_1, *self.middle, self.output)
SyntaxError: only named arguments may follow *expression

Tensors have different dimensions

Hi,
firstly, thank you for your work on this repo. I tried to run your code, but I get this exception:

Traceback (most recent call last):
  File "/home/lukas/dev/pytorch-yolo-v3/detect.py", line 183, in <module>
    prediction = write_results(prediction, confidence, num_classes, nms = True, nms_conf = nms_thesh)
  File "/home/lukas/dev/pytorch-yolo-v3/util.py", line 189, in write_results
    output = torch.cat(seq, 0)
RuntimeError: invalid argument 0: Tensors must have same number of dimensions: got 1 and 2 at /tmp/pip-w41ywlv_-build/aten/src/THC/generic/THCTensorMath.cu:102

I tried to debug what is happening in the code, but it is not very clear to me. I only noticed that in my case, one tensor's size is 7 and the other is (7,1). I am also adding a screen.
image

So I have modified the 7 sized tensor into (7,1) dimension but then I get another exception from further code.

Traceback (most recent call last):
  File "/opt/pycharm-2018.1/helpers/pydev/pydevd.py", line 1664, in <module>
    main()
  File "/opt/pycharm-2018.1/helpers/pydev/pydevd.py", line 1658, in main
    globals = debugger.run(setup['file'], None, None, is_module)
  File "/opt/pycharm-2018.1/helpers/pydev/pydevd.py", line 1068, in run
    pydev_imports.execfile(file, globals, locals)  # execute the script
  File "/opt/pycharm-2018.1/helpers/pydev/_pydev_imps/_pydev_execfile.py", line 18, in execfile
    exec(compile(contents+"\n", file, 'exec'), glob, loc)
  File "/home/lukas/dev/pytorch-yolo-v3/detect.py", line 213, in <module>
    objs = [classes[int(x[-1])] for x in output if int(x[0]) == im_id]
  File "/home/lukas/dev/pytorch-yolo-v3/detect.py", line 213, in <listcomp>
    objs = [classes[int(x[-1])] for x in output if int(x[0]) == im_id]
IndexError: list index out of range

Don't you please know what I am doing wrong? It is my first time working with PyTorch, so I am not very experienced and I don't know how to fix it myself.

Using Ubuntu 16.4, Python 3.6, CUDA 9.0, PyTorch 0.4.
Thanks

bounding box coordinate

Hi @ayooshkathuria ,
in the detect.py, I see that line 272 outputs the bounding box coordinate for every object detected.
what exactly is the order of the coordinate? is it [image width height x-center y center prob prob class]?

Thanks

letter box image

Hi, I just need to use letter box function from the code. It is giving me error as I am confused what would be the inp_dim . Is it the shape of input image ? or any other value. It is giving error bothways

Cannot capture source

Hi,I`m try to use your code to run the video ,my video name is chicken.avi.But when I code python video_demo.py --video chicken.avi,the result is
Traceback (most recent call last):
File "video_demo.py", line 119, in
assert cap.isOpened(), 'Cannot capture source'
AssertionError: Cannot capture source
Is it means that I have the wrong code or other question?Can anybody help me?Please!!

function create_modules() bug

the darknet.py line 209:
x["layers"] = x["layers"].split(',')
is not right.
if you print this function output you will get:
---> 59 x["layers"] = x["layers"].split(",")
AttributeError: 'list' object has no attribute 'split'

if split x["layers"], how to do following code:
start = int(x["layers"][0])
and
end = int(x["layers"][1])
no [0],[1] index.
I think this line should be delete, however, if delete this line, run thiese follow commond:
blocks = parse_cfg("cfg/yolov3.cfg")
print(create_modules(blocks))
the output gets error as well.
---> 61 start = int(x["layers"][0])
62 #end, if there exists one.
63 try:

ValueError: invalid literal for int() with base 10: '-'
can somebody help

no outputs

S E:\condaDev\pytorch-yolo-v3-master> python detect.py --images imgs --det det
Loading network.....
Network successfully loaded
C:\Users\Max\Anaconda3\envs\Pytorch\lib\site-packages\torch\nn\modules\upsampling.py:122: UserWarning: nn.Upsampling is deprecated. Use nn.functional.interpolate instead.
warnings.warn("nn.Upsampling is deprecated. Use nn.functional.interpolate instead.")
E:\condaDev\pytorch-yolo-v3-master\imgs\dog.jpg predicted in 1.411 seconds
Objects Detected: bicycle truck dog

E:\condaDev\pytorch-yolo-v3-master\imgs\eagle.jpg predicted in 1.386 seconds
Objects Detected: bird

E:\condaDev\pytorch-yolo-v3-master\imgs\giraffe.jpg predicted in 1.414 seconds
Objects Detected: zebra giraffe giraffe

E:\condaDev\pytorch-yolo-v3-master\imgs\herd_of_horses.jpg predicted in 1.420 seconds
Objects Detected: horse horse horse horse

E:\condaDev\pytorch-yolo-v3-master\imgs\img1.jpg predicted in 1.406 seconds
Objects Detected: person dog

E:\condaDev\pytorch-yolo-v3-master\imgs\img2.jpg predicted in 1.410 seconds
Objects Detected: train

E:\condaDev\pytorch-yolo-v3-master\imgs\img3.jpg predicted in 1.407 seconds
Objects Detected: car car car car car car car truck traffic light

E:\condaDev\pytorch-yolo-v3-master\imgs\img4.jpg predicted in 1.402 seconds
Objects Detected: chair chair chair clock

E:\condaDev\pytorch-yolo-v3-master\imgs\messi.jpg predicted in 1.401 seconds
Objects Detected: person person person sports ball

E:\condaDev\pytorch-yolo-v3-master\imgs\person.jpg predicted in 1.542 seconds
Objects Detected: person dog horse

SUMMARY

Task : Time Taken (in seconds)

Reading addresses : 0.000
Loading batch : 1.614
Detection (11 images) : 15.668
Output Processing : 0.000
Drawing Boxes : 0.005
Average time_per_img : 1.571

no outputs found in det folder

Number of Images inference

Hi,
Thank you for your work. It is really easier to use pytorch instead of darknet.
Nevertheless, I have tried to run the command on a folder with thousands of images
python3 detect.py --images img --det det --reso 320 --weights yolov3.weights
And got a killed error:

Loading network.....
Network successfully loaded
Killed

I don't know why it is working for few images and not for more. I have tried to use a batch size of 1 image but I still get the same error.
Thanks in advance

only the first row of data in the names file showed

Hello, I have a little problem. I set count=4, the model used is tiny-yolo, I have 6 kinds of data. When I was testing, only the first row of data in the names file showed, why? My cfg file shown below:
`
[net]

Testing

#batch=1
#subdivisions=1

Training

batch=1
subdivisions=1
width=416
height=416
channels=3
momentum=0.9
decay=0.0005
angle=0
saturation = 1.5
exposure = 1.5
hue=.1

learning_rate=0.00001
burn_in=1000
max_batches = 500200
policy=steps
steps=400000,450000
scales=.1,.1

[convolutional]
batch_normalize=1
filters=16
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=32
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=64
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=128
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=256
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=512
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=1

[convolutional]
batch_normalize=1
filters=1024
size=3
stride=1
pad=1
activation=leaky

###########

[convolutional]
batch_normalize=1
filters=256
size=1
stride=1
pad=1
activation=leaky

[convolutional]
batch_normalize=1
filters=512
size=3
stride=1
pad=1
activation=leaky

[convolutional]
size=1
stride=1
pad=1
filters=33
activation=linear

[yolo]
mask = 3,4,5
anchors = 10,14, 23,27, 37,58, 81,82, 135,169, 344,319
classes=6
num=6
jitter=.3
ignore_thresh = .7
truth_thresh = 1
random=1

[route]
layers = -4

[convolutional]
batch_normalize=1
filters=128
size=1
stride=1
pad=1
activation=leaky

[upsample]
stride=2

[route]
layers = -1, 8

[convolutional]
batch_normalize=1
filters=256
size=3
stride=1
pad=1
activation=leaky

[convolutional]
size=1
stride=1
pad=1
filters=33
activation=linear

[yolo]
mask = 0,1,2
anchors = 10,14, 23,27, 37,58, 81,82, 135,169, 344,319
classes=6
num=6
jitter=.3
ignore_thresh = .7
truth_thresh = 1
random=1
small_object=1
`

The test results are shown below :
1-NG.bmp predicted in 0.112 seconds
Objects Detected: ok ok ok ok ok ok ok ok ok ok ok ok ok ok

Please help me,thanks a lot

Problem when running video_demo_half.py

At first, I found the video_demo.py support mp4 video surprisingly.
But when I run the video_demo_half.py, there is some mistake happened
RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.cuda.HalfTensor) should be the same
Maybe there is another weights of 16-bit which is not released?

RuntimeError: invalid argument 2: size '[1 x 18 x 1369]' is invalid for input with 23328 elements at ..\aten\src\TH\THStorage.cpp:84

I am stuck in this error please help...

2018-09-26 12:39:49 epoch 2, processed 200 samples, lr 0.000010
torch.Size([1, 18, 18, 18])
D:\Softwares\anacond33\lib\site-packages\torch\nn\modules\upsampling.py:122: UserWarning: nn.Upsampling is deprecated. Use nn.functional.interpolate instead.
warnings.warn("nn.Upsampling is deprecated. Use nn.functional.interpolate instead.")
torch.Size([1, 18, 36, 36])

RuntimeError Traceback (most recent call last)
in ()
----> 1 train(2)

in train(epoch)
42
43
---> 44 output = model(data)
45
46

D:\Softwares\anacond33\lib\site-packages\torch\nn\modules\module.py in call(self, *input, **kwargs)
475 result = self._slow_forward(*input, **kwargs)
476 else:
--> 477 result = self.forward(*input, **kwargs)
478 for hook in self._forward_hooks.values():
479 hook_result = hook(self, input, result)

D:\Softwares\anacond33\lib\site-packages\torch\nn\parallel\data_parallel.py in forward(self, *inputs, **kwargs)
119 inputs, kwargs = self.scatter(inputs, kwargs, self.device_ids)
120 if len(self.device_ids) == 1:
--> 121 return self.module(*inputs[0], **kwargs[0])
122 replicas = self.replicate(self.module, self.device_ids[:len(inputs)])
123 outputs = self.parallel_apply(replicas, inputs, kwargs)

D:\Softwares\anacond33\lib\site-packages\torch\nn\modules\module.py in call(self, *input, **kwargs)
475 result = self._slow_forward(*input, **kwargs)
476 else:
--> 477 result = self.forward(*input, **kwargs)
478 for hook in self._forward_hooks.values():
479 hook_result = hook(self, input, result)

~\Aandom\yolo_pytorch\pytorch_yolo_darknet_arc.py in forward(self, x, CUDA)
386 x = x.data
387 #print(x.shape)
--> 388 x = predict_transform(x, inp_dim, anchors, num_classes, CUDA)
389
390 if type(x) == int:

~\Aandom\yolo_pytorch\utils.py in predict_transform(prediction, inp_dim, anchors, num_classes, CUDA)
72
73 print(prediction.shape)
---> 74 prediction = prediction.view(batch_size, bbox_attrsnum_anchors, grid_sizegrid_size)
75 prediction = prediction.transpose(1,2).contiguous()
76 prediction = prediction.view(batch_size, grid_sizegrid_sizenum_anchors, bbox_attrs)

RuntimeError: invalid argument 2: size '[1 x 18 x 1369]' is invalid for input with 23328 elements at ..\aten\src\TH\THStorage.cpp:84

Refactor for usability & maintainability

Love what you've done here (and especially the in-depth tutorial on the Paperspace blog!). I've been looking for a performant pytorch implementation of YOLOv3 and this seems to be the closest thing that fits the bill.

One suggestion: Right now the repo feels more like a proof of concept than a well organized library (one example: the simple task of figuring out how to plug in an image and get a list of predictions out to use in another python app isn't super straightforward).

Would love to see this progress towards a more refined library (and if I have some time I'll certainly try to help out with any refactoring I can).

Thanks for all your work on this!

error in Different batch size in detect.py

It seems the code can not work when I make a change to the batch size of detect.py

Traceback (most recent call last):
  File "/home/dong/PycharmProjects/SIMD/pytorch-yolo-v3/detect.py", line 200, in <module>
    prediction = write_results(prediction, confidence, num_classes, nms = True, nms_conf = nms_thesh)
  File "/home/dong/PycharmProjects/SIMD/pytorch-yolo-v3/util.py", line 155, in write_results
    img_classes = unique(image_pred_[:,-1])
IndexError: too many indices for tensor of dimension 1

How to load pre-trained model

Hi, I do like your code which help me understand the YOLO framework. But it is regretful that there are no training process. Now I'm trying to write the training codes by myself but I don't know how to load the pre-trained model given by the official YOLO authors darknet53.conv.74. I tried straightly use the load_weights() method of Darknet but failed. The error reports as RuntimeEroor: the given numpy array has zero-sized dimensions. Zero-sized dimensions are not supported in Pytorch. I guess this error happens when loading step encounter BN layer. Can you help me with this? Thanks.

Training

Thanks for creating this, it looks great! I would love to use it, but I would need the possibility to train according to my needs - can you probably give an estimate when training will be available?

Thanks and kind regards
Ernst

ValueError: not enough values to unpack (expected 2, got 1)

hi, I got a trouble during the tutorial chapter 2.
when I did this part,

Testing the code

You can test your code by typing the following lines at the end of darknet.py and running the file.

blocks = parse_cfg("cfg/yolov3.cfg")
print(create_modules(blocks))

there's an error occured.

File "darknet.py", line 147, in
blocks = parse_cfg("cfg/yolov3.cfg")
File "darknet.py", line 26, in parse_cfg
key, value = line.split("=")
ValueError: not enough values to unpack (expected 2, got 1)

REALLY NEED HELP, thanks

Sorry, I am a rookie, I downloaded your program, this problem occurs after running. Ask you for some question.Thank you very match!

Traceback (most recent call last):
File "D:/deep_learn/pytorch-yolo-v3-master/pytorch-yolo-v3-master/video_demo_half.py", line 100, in
model.load_weights(args.weightsfile)
File "D:\deep_learn\pytorch-yolo-v3-master\pytorch-yolo-v3-master\darknet.py", line 367, in load_weights
fp = open(weightfile, "rb")
FileNotFoundError: [Errno 2] No such file or directory: 'yolov3.weights'

Images are labelled in compressed clusters when using own cfg and weights

Hi,

After I successfully loaded my own yolov3-voc-own.weights obtained from training the network with Pascal VOC, referred to the yolov3-voc.cfg I used for training file, and modified classes names and number (from coco.names to voc.names and from 80 to 20) I was surprised at the results.

Please have a look at the following images
Herd of horses: https://ibb.co/iO8WFo
Messi/Inista: https://ibb.co/nMNaao
Guy with poncho and dog: https://ibb.co/j5bU1T

In all of them, there is a pattern of identifying the right combination of images (incorrectly, but clearly finding something) but in the wrong location, and condensed. My guess is that the recognition happens in one of the 'zoomed' version of the image, and it's identified anchors are not correctly remapped to the original size of the image.

Is this any reasonable? Do you think this could be a bug from the 'interpretation' of the .cfg file? Or perhaps I'm missing another piece of the puzzle?

Another possible point of code that could possibly be related to this is the input argument of scales, where I feed 1,2,3 as default. Could it be that it should be matched with some properties of my network?

Thanks in advance!

Calling the model object?

Working on making your detect.py a reusable module, trying to work on having the script treat images passed as single images instead of lists of images from a directory. Going well so far, issue I'm on right now, what is this doing?

with torch.no_grad(): prediction = self.model(Variable(batch), self.CUDA)

What I'm confused about is what self.model(Variable(batch), self.CUDA) is doing. It looks like you're calling the self.model object, but i can't see what that does within the Darknet class that defines it.

path consist of "\\" can cause bug

Hi

In the line 302, det_names = pd.Series(imlist).apply(lambda x: "{}/det_{}".format(args.det,x.split("/")[-1]))

this can be problem. In my case, it doesn't split path, because of "/".

but i works well by changing "/" to "\"

Please note this issue

train code

@ayooshkathuria , hi, thanks a lot for your work. And I am waiting for the train code. Could you tell me when the train code is available, plz? Because I want to be lazy. Hope this quest not bother you.

yolov3-tiny doesnt work

i have the cfg and weights file. I tried to run, but it gives this error

File "object_detection/darknet.py", line 323, in forward
x = torch.cat((map1, map2), 1)
RuntimeError: invalid argument 0: Sizes of tensors must match except in dimension 1. Got 416 and 832 in dimension 2 at /pytorch/aten/src/THC/generic/THCTensorMath.cu:111

License missing

Hi, what's the license of this code? Is it under MIT as pytorch-yolo2 by marvis?

Error when trying to convert darknet.py to ONNX

Hi!

I'm trying to create a script to convert the darknet.py Pytorch model to ONNX model using CPU based on this example:

# onnx-convert.py

from darknet import Darknet

import torch 
import torch.onnx
from torch.autograd import Variable
import torchvision

cfgfile = './cfg/yolov3.cfg'
weightfile =  './yolov3.weights'
imgfile = './imgs/dog.jpg'
resolution = 416

model = Darknet(cfgfile)
model.load_weights(weightfile)

dummy_input = Variable(torch.randn(1, 3, resolution, resolution))
torch.onnx.export(model, dummy_input, "yolo.onnx")

But when I run python onnx-convert.py, the following error happens:

TypeError: forward() missing 1 required positional argument: 'CUDA'

I changed the Darknet#forward(self, x, CUDA) member signature to Darknet#forward(self, x, CUDA = False). Now the following error happens:

RuntimeError: invalid argument 2: size '[1 x 255 x 2809]' is invalid for input with 689520 elements at /pytorch/aten/src/TH/THStorage.c:41

I changed my script to include following condition:

from darknet import Darknet

...
model = Darknet(cfgfile)
model.net_info["height"] = resolution
...
torch.onnx.export(model, dummy_input, "yolo.onnx")

The following error happens:
RuntimeError: /pytorch/torch/csrc/jit/tracer.h:120: getTracingState: Assertion state failed.

Someone can help me?

Train branch

Hi,
What is the status of the train branch ? I'd be down for contributing if you wish ?

Test questions

Why can you use your code to perform detection in the official weight file, while running the weight file that I trained myself can not detect it?

Have problem running the demo.py

I run the code on a Linux server with Pytorch 0.4.0. And I get the following error:
qq 20180427212738

I use the py4 branch. Please tell me how to run this.

output issues between windows and linux

It totally works in linux, output images saved in det folder.
However, no output images saved in det folder on windows, even it detects objects in images.
it runs correctly in detection part, but save output images into folder failed.
How to solve this problem?

Error when loading own trained weight file

Hi,

I tested loading my own trained weights which I got using darknet's official guide. Basically, I substituted yolov3.weights with my own yolov3-own.weights as the default --weights argument in the code.

This is console's output:

Loading network.....
Traceback (most recent call last):

  File "<ipython-input-3-1d1acba7b8a9>", line 1, in <module>
    runfile('/home/raggot/Projects/test/detect.py', wdir='/home/raggot/Projects/test')

  File "/usr/local/lib/python3.5/dist-packages/spyder/utils/site/sitecustomize.py", line 705, in runfile
    execfile(filename, namespace)

  File "/usr/local/lib/python3.5/dist-packages/spyder/utils/site/sitecustomize.py", line 102, in execfile
    exec(compile(f.read(), filename, 'exec'), namespace)

  File "/home/raggot/Projects/test/detect.py", line 89, in <module>
    model.load_weights(args.weightsfile)

  File "/home/raggot/Projects/test/darknet.py", line 448, in load_weights
    conv_weights = conv_weights.view_as(conv.weight.data)

  File "/usr/local/lib/python3.5/dist-packages/torch/tensor.py", line 230, in view_as
    return self.view(tensor.size())

RuntimeError: invalid argument 2: size '[256 x 128 x 3 x 3]' is invalid for input with 37347 elements at /pytorch/aten/src/TH/THStorage.c:41

What is the problem here? What are the 37347 elements the error refers to?

Thanks.

Issue with printing on windows

Hey I ran the detect.py script on windows machine and no images were being output. I found the error, it's on line 302 of detect.py:

det_names = pd.Series(imlist).apply(lambda x: "{}/det_{}".format(args.det,x.split("/")[-1]))

This works for linux and mac machines but not on windows, it's an easy fix though. I tried to push a PR but didnt realize I couldnt. The fix is simple, add import platform to the top of detect.py and change line 302 to

if platform.system() == 'Windows':
        det_names = pd.Series(imlist).apply(lambda x: "{}/det_{}".format(args.det,x.split("\\")[-1]))
    else:
        det_names = pd.Series(imlist).apply(lambda x: "{}/det_{}".format(args.det,x.split("/")[-1]))```

RuntimeError: invalid argument 2: size '[18 x 256 x 1 x 1]' is invalid for input with 4607 elements

Hi, I'm trying to use your code to do some detection work.
I'm successful while using yolov3.cfg and yolov3.weights.
But when I change this cfg file and weights to my detection, error occurs.
I only change the classes num to 1 and the error is here:

Traceback (most recent call last):
  File "D:/GitHub/Graduation/YOLO_v3_self/detect.py", line 109, in <module>
    model.load_weights(args.weightsfile)
  File "D:\GitHub\Graduation\YOLO_v3_self\darknet.py", line 425, in load_weights
    conv_weights = conv_weights.view_as(conv.weight.data)
  File "C:\software\Anaconda3\envs\my_dev\lib\site-packages\torch\tensor.py", line 230, in view_as
    return self.view(tensor.size())
RuntimeError: invalid argument 2: size '[18 x 256 x 1 x 1]' is invalid for input with 4607 elements at ..\src\TH\THStorage.c:41

Is there some solution?
Thanks!

model predict question

Hello, the model I trained with Darknet can't detect objects when running on your code. But the weight file given by official Darknet can be detected. What's the reason?

error in darknet.py

anchors = x["anchors"].split(",")
TypeError: string indices must be integers, not str

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.