You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I tried to extract 3D features following the three-step pipeline using blip_sam.py file
There are a few questions about the details:
In line 53-54 of second_step/blip_sam.py, you have:
raw_image = cv2.imread(INPUT_IMAGE_PATH)
raw_image = cv2.resize(raw_image, (512, 512))
is it correct to go without using: img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)?
Once I generated my own point cloud features, how could I verify if the features are consistent to yours? Could you provide a subset of the features (eg. scannet)? I also tried to compute the similarity between point cloud features with text description. However, since the 1408 dim features are the hidden layer output, it is not feasible to compute the similarity with text features which are 256 dim. Do you have any suggestions on this?
The text was updated successfully, but these errors were encountered:
Hi, I tried to extract 3D features following the three-step pipeline using blip_sam.py file
There are a few questions about the details:
In line 53-54 of second_step/blip_sam.py, you have:
raw_image = cv2.imread(INPUT_IMAGE_PATH)
raw_image = cv2.resize(raw_image, (512, 512))
is it correct to go without using: img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)?
Once I generated my own point cloud features, how could I verify if the features are consistent to yours? Could you provide a subset of the features (eg. scannet)? I also tried to compute the similarity between point cloud features with text description. However, since the 1408 dim features are the hidden layer output, it is not feasible to compute the similarity with text features which are 256 dim. Do you have any suggestions on this?
The text was updated successfully, but these errors were encountered: